+ source /setup_export_logs.sh ++ CLICKHOUSE_CI_LOGS_CREDENTIALS=/tmp/export-logs-config.sh ++ CLICKHOUSE_CI_LOGS_USER=ci ++ CLICKHOUSE_CI_LOGS_CLUSTER=system_logs_export ++ EXTRA_COLUMNS='pull_request_number UInt32, commit_sha String, check_start_time DateTime('\''UTC'\''), check_name LowCardinality(String), instance_type LowCardinality(String), instance_id String, INDEX ix_pr (pull_request_number) TYPE set(100), INDEX ix_commit (commit_sha) TYPE set(100), INDEX ix_check_time (check_start_time) TYPE minmax, ' ++ EXTRA_COLUMNS_EXPRESSION='CAST(0 AS UInt32) AS pull_request_number, '\'''\'' AS commit_sha, now() AS check_start_time, toLowCardinality('\'''\'') AS check_name, toLowCardinality('\'''\'') AS instance_type, '\'''\'' AS instance_id' ++ EXTRA_ORDER_BY_COLUMNS=check_name ++ EXTRA_COLUMNS_TRACE_LOG='pull_request_number UInt32, commit_sha String, check_start_time DateTime('\''UTC'\''), check_name LowCardinality(String), instance_type LowCardinality(String), instance_id String, INDEX ix_pr (pull_request_number) TYPE set(100), INDEX ix_commit (commit_sha) TYPE set(100), INDEX ix_check_time (check_start_time) TYPE minmax, symbols Array(LowCardinality(String)), lines Array(LowCardinality(String)), ' ++ EXTRA_COLUMNS_EXPRESSION_TRACE_LOG='CAST(0 AS UInt32) AS pull_request_number, '\'''\'' AS commit_sha, now() AS check_start_time, toLowCardinality('\'''\'') AS check_name, toLowCardinality('\'''\'') AS instance_type, '\'''\'' AS instance_id, arrayMap(x -> demangle(addressToSymbol(x)), trace)::Array(LowCardinality(String)) AS symbols, arrayMap(x -> addressToLine(x), trace)::Array(LowCardinality(String)) AS lines' ++ EXTRA_COLUMNS_COVERAGE_LOG='pull_request_number UInt32, commit_sha String, check_start_time DateTime('\''UTC'\''), check_name LowCardinality(String), instance_type LowCardinality(String), instance_id String, INDEX ix_pr (pull_request_number) TYPE set(100), INDEX ix_commit (commit_sha) TYPE set(100), INDEX ix_check_time (check_start_time) TYPE minmax, symbols Array(LowCardinality(String)), ' ++ EXTRA_COLUMNS_EXPRESSION_COVERAGE_LOG='CAST(0 AS UInt32) AS pull_request_number, '\'''\'' AS commit_sha, now() AS check_start_time, toLowCardinality('\'''\'') AS check_name, toLowCardinality('\'''\'') AS instance_type, '\'''\'' AS instance_id, arrayDistinct(arrayMap(x -> demangle(addressToSymbol(x)), coverage))::Array(LowCardinality(String)) AS symbols' + source /repo/tests/docker_scripts/stress_tests.lib ++ OK='\tOK\t\N\t' ++ FAIL='\tFAIL\t\N\t' ++ FAILURE_CONTEXT_LINES=100 ++ FAILURE_CONTEXT_MAX_LINE_WIDTH=300 + dmesg --clear + set -e -x -a + USE_DATABASE_REPLICATED=0 + USE_SHARED_CATALOG=0 ++ rg -v '#' /usr/share/zoneinfo/zone.tab ++ awk '{print $3}' ++ shuf ++ head -n1 + TZ=America/St_Lucia Chosen random timezone America/St_Lucia + echo 'Chosen random timezone America/St_Lucia' + ln -snf /usr/share/zoneinfo/America/St_Lucia /etc/localtime + echo America/St_Lucia + dpkg -i package_folder/clickhouse-common-static_24.12.2.20221.altinityantalya+tsan_amd64.deb Selecting previously unselected package clickhouse-common-static. (Reading database ... 48425 files and directories currently installed.) Preparing to unpack .../clickhouse-common-static_24.12.2.20221.altinityantalya+tsan_amd64.deb ... Unpacking clickhouse-common-static (24.12.2.20221.altinityantalya+tsan) ... Setting up clickhouse-common-static (24.12.2.20221.altinityantalya+tsan) ... + dpkg -i package_folder/clickhouse-common-static-dbg_24.12.2.20221.altinityantalya+tsan_amd64.deb Selecting previously unselected package clickhouse-common-static-dbg. (Reading database ... 48452 files and directories currently installed.) Preparing to unpack .../clickhouse-common-static-dbg_24.12.2.20221.altinityantalya+tsan_amd64.deb ... Unpacking clickhouse-common-static-dbg (24.12.2.20221.altinityantalya+tsan) ... Setting up clickhouse-common-static-dbg (24.12.2.20221.altinityantalya+tsan) ... + dpkg -i package_folder/clickhouse-odbc-bridge_24.12.2.20221.altinityantalya+tsan_amd64.deb Selecting previously unselected package clickhouse-odbc-bridge. (Reading database ... 48459 files and directories currently installed.) Preparing to unpack .../clickhouse-odbc-bridge_24.12.2.20221.altinityantalya+tsan_amd64.deb ... Unpacking clickhouse-odbc-bridge (24.12.2.20221.altinityantalya+tsan) ... Setting up clickhouse-odbc-bridge (24.12.2.20221.altinityantalya+tsan) ... + dpkg -i package_folder/clickhouse-library-bridge_24.12.2.20221.altinityantalya+tsan_amd64.deb Selecting previously unselected package clickhouse-library-bridge. (Reading database ... 48465 files and directories currently installed.) Preparing to unpack .../clickhouse-library-bridge_24.12.2.20221.altinityantalya+tsan_amd64.deb ... Unpacking clickhouse-library-bridge (24.12.2.20221.altinityantalya+tsan) ... Setting up clickhouse-library-bridge (24.12.2.20221.altinityantalya+tsan) ... + dpkg -i package_folder/clickhouse-server_24.12.2.20221.altinityantalya+tsan_amd64.deb Selecting previously unselected package clickhouse-server. (Reading database ... 48471 files and directories currently installed.) Preparing to unpack .../clickhouse-server_24.12.2.20221.altinityantalya+tsan_amd64.deb ... Unpacking clickhouse-server (24.12.2.20221.altinityantalya+tsan) ... Setting up clickhouse-server (24.12.2.20221.altinityantalya+tsan) ... ClickHouse binary is already located at /usr/bin/clickhouse Symlink /usr/bin/clickhouse-server already exists but it points to /clickhouse. Will replace the old symlink to /usr/bin/clickhouse. Creating symlink /usr/bin/clickhouse-server to /usr/bin/clickhouse. Creating symlink /usr/bin/clickhouse-client to /usr/bin/clickhouse. Creating symlink /usr/bin/clickhouse-local to /usr/bin/clickhouse. Creating symlink /usr/bin/clickhouse-benchmark to /usr/bin/clickhouse. Creating symlink /usr/bin/clickhouse-obfuscator to /usr/bin/clickhouse. Creating symlink /usr/bin/clickhouse-git-import to /usr/bin/clickhouse. Creating symlink /usr/bin/clickhouse-compressor to /usr/bin/clickhouse. Creating symlink /usr/bin/clickhouse-format to /usr/bin/clickhouse. Symlink /usr/bin/clickhouse-extract-from-config already exists but it points to /clickhouse. Will replace the old symlink to /usr/bin/clickhouse. Creating symlink /usr/bin/clickhouse-extract-from-config to /usr/bin/clickhouse. Symlink /usr/bin/clickhouse-keeper already exists but it points to /clickhouse. Will replace the old symlink to /usr/bin/clickhouse. Creating symlink /usr/bin/clickhouse-keeper to /usr/bin/clickhouse. Symlink /usr/bin/clickhouse-keeper-converter already exists but it points to /clickhouse. Will replace the old symlink to /usr/bin/clickhouse. Creating symlink /usr/bin/clickhouse-keeper-converter to /usr/bin/clickhouse. Creating symlink /usr/bin/clickhouse-disks to /usr/bin/clickhouse. Creating symlink /usr/bin/ch to /usr/bin/clickhouse. Creating symlink /usr/bin/chl to /usr/bin/clickhouse. Creating symlink /usr/bin/chc to /usr/bin/clickhouse. Creating clickhouse group if it does not exist. groupadd -r clickhouse Creating clickhouse user if it does not exist. useradd -r --shell /bin/false --home-dir /nonexistent -g clickhouse clickhouse Will set ulimits for clickhouse user in /etc/security/limits.d/clickhouse.conf. Creating config directory /etc/clickhouse-server/config.d that is used for tweaks of main server configuration. Creating config directory /etc/clickhouse-server/users.d that is used for tweaks of users configuration. Config file /etc/clickhouse-server/config.xml already exists, will keep it and extract path info from it. /etc/clickhouse-server/config.xml has /var/lib/clickhouse/ as data path. /etc/clickhouse-server/config.xml has /var/log/clickhouse-server/ as log path. Users config file /etc/clickhouse-server/users.xml already exists, will keep it and extract users info from it. Log directory /var/log/clickhouse-server/ already exists. Creating data directory /var/lib/clickhouse/. Creating pid directory /var/run/clickhouse-server. chown -R clickhouse:clickhouse '/var/log/clickhouse-server/' chown -R clickhouse:clickhouse '/var/run/clickhouse-server' chown clickhouse:clickhouse '/var/lib/clickhouse/' groupadd -r clickhouse-bridge useradd -r --shell /bin/false --home-dir /nonexistent -g clickhouse-bridge clickhouse-bridge chown -R clickhouse-bridge:clickhouse-bridge '/usr/bin/clickhouse-odbc-bridge' chown -R clickhouse-bridge:clickhouse-bridge '/usr/bin/clickhouse-library-bridge' Password for the default user is an empty string. See /etc/clickhouse-server/users.xml and /etc/clickhouse-server/users.d to change it. Setting capabilities for clickhouse binary. This is optional. chown -R clickhouse:clickhouse '/etc/clickhouse-server' ClickHouse has been successfully installed. Start clickhouse-server with: sudo clickhouse start Start clickhouse-client with: clickhouse-client + dpkg -i package_folder/clickhouse-client_24.12.2.20221.altinityantalya+tsan_amd64.deb Selecting previously unselected package clickhouse-client. (Reading database ... 48488 files and directories currently installed.) Preparing to unpack .../clickhouse-client_24.12.2.20221.altinityantalya+tsan_amd64.deb ... Unpacking clickhouse-client (24.12.2.20221.altinityantalya+tsan) ... Setting up clickhouse-client (24.12.2.20221.altinityantalya+tsan) ... + echo '' + [[ -z '' ]] + ch --query 'SELECT 1' 1 + chl --query 'SELECT 1' 1 + chc --version ClickHouse client version 24.12.2.20221.altinityantalya (altinity build). + ln -sf /repo/tests/clickhouse-test /usr/bin/clickhouse-test + export CLICKHOUSE_GRPC_CLIENT=/repo/utils/grpc-client/clickhouse-grpc-client.py + CLICKHOUSE_GRPC_CLIENT=/repo/utils/grpc-client/clickhouse-grpc-client.py + source /repo/tests/docker_scripts/attach_gdb.lib ++ source /repo/tests/docker_scripts/utils.lib +++ sysctl kernel.core_pattern=core.%e.%p-%P kernel.core_pattern = core.%e.%p-%P +++ sysctl fs.suid_dumpable=1 fs.suid_dumpable = 1 + source /repo/tests/docker_scripts/utils.lib ++ sysctl kernel.core_pattern=core.%e.%p-%P kernel.core_pattern = core.%e.%p-%P ++ sysctl fs.suid_dumpable=1 fs.suid_dumpable = 1 + /repo/tests/config/install.sh + DEST_SERVER_PATH=/etc/clickhouse-server + DEST_CLIENT_PATH=/etc/clickhouse-client +++ dirname /repo/tests/config/install.sh ++ cd /repo/tests/config ++ pwd -P + SRC_PATH=/repo/tests/config + '[' 0 -ge 2 ']' + FAST_TEST=0 + NO_AZURE=0 + [[ 0 -gt 0 ]] + echo 'Going to install test configs from /repo/tests/config into /etc/clickhouse-server' + mkdir -p /etc/clickhouse-server/config.d/ Going to install test configs from /repo/tests/config into /etc/clickhouse-server + mkdir -p /etc/clickhouse-server/users.d/ + mkdir -p /etc/clickhouse-client + ln -sf /repo/tests/config/config.d/zookeeper_write.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/max_num_to_warn.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/listen.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/text_log.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/blob_storage_log.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/custom_settings_prefixes.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/database_catalog_drop_table_concurrency.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/enable_access_control_improvements.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/macros.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/secure_ports.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/clusters.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/graphite.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/graphite_alternative.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/grpc_protocol.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/database_atomic.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/max_concurrent_queries.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/merge_tree_settings.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/backoff_failed_mutation.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/merge_tree_old_dirs_cleanup.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/test_cluster_with_incorrect_pw.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/keeper_port.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/logging_no_rotate.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/merge_tree.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/lost_forever_check.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/tcp_with_proxy.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/prometheus.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/top_level_domains_lists.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/top_level_domains_path.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/transactions.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/encryption.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/CORS.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/zookeeper_log.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/logger_trace.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/named_collection.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/ssl_certs.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/filesystem_cache_log.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/session_log.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/system_unfreeze.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/enable_zero_copy_replication.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/nlp.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/forbidden_headers.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/enable_keeper_map.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/custom_disks_base_path.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/display_name.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/compressed_marks_and_index.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/disable_s3_env_credentials.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/enable_wait_for_shutdown_replicated_tables.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/backups.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/filesystem_caches_path.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/validate_tcp_client_information.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/zero_copy_destructive_operations.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/block_number.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/handlers.yaml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/serverwide_trace_collector.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/rocksdb.xml /etc/clickhouse-server/config.d/ + '[' 0 '!=' 1 ']' + ln -sf /repo/tests/config/config.d/legacy_geobase.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/users.d/log_queries.xml /etc/clickhouse-server/users.d/ + ln -sf /repo/tests/config/users.d/readonly.xml /etc/clickhouse-server/users.d/ + ln -sf /repo/tests/config/users.d/access_management.xml /etc/clickhouse-server/users.d/ + ln -sf /repo/tests/config/users.d/database_atomic_drop_detach_sync.xml /etc/clickhouse-server/users.d/ + ln -sf /repo/tests/config/users.d/opentelemetry.xml /etc/clickhouse-server/users.d/ + ln -sf /repo/tests/config/users.d/remote_queries.xml /etc/clickhouse-server/users.d/ + ln -sf /repo/tests/config/users.d/session_log_test.xml /etc/clickhouse-server/users.d/ + ln -sf /repo/tests/config/users.d/memory_profiler.xml /etc/clickhouse-server/users.d/ + ln -sf /repo/tests/config/users.d/no_fsync_metadata.xml /etc/clickhouse-server/users.d/ + ln -sf /repo/tests/config/users.d/filelog.xml /etc/clickhouse-server/users.d/ + ln -sf /repo/tests/config/users.d/enable_blobs_check.xml /etc/clickhouse-server/users.d/ + ln -sf /repo/tests/config/users.d/marks.xml /etc/clickhouse-server/users.d/ + ln -sf /repo/tests/config/users.d/insert_keeper_retries.xml /etc/clickhouse-server/users.d/ + ln -sf /repo/tests/config/users.d/prefetch_settings.xml /etc/clickhouse-server/users.d/ + ln -sf /repo/tests/config/users.d/nonconst_timezone.xml /etc/clickhouse-server/users.d/ + ln -sf /repo/tests/config/users.d/allow_introspection_functions.yaml /etc/clickhouse-server/users.d/ + ln -sf /repo/tests/config/users.d/replicated_ddl_entry.xml /etc/clickhouse-server/users.d/ + ln -sf /repo/tests/config/users.d/limits.yaml /etc/clickhouse-server/users.d/ + [[ -n '' ]] + ln -sf /repo/tests/config/users.d/timeouts.xml /etc/clickhouse-server/users.d/ + ln -sf /repo/tests/config/ints_dictionary.xml /etc/clickhouse-server/ + ln -sf /repo/tests/config/strings_dictionary.xml /etc/clickhouse-server/ + ln -sf /repo/tests/config/decimals_dictionary.xml /etc/clickhouse-server/ + ln -sf /repo/tests/config/executable_dictionary.xml /etc/clickhouse-server/ + ln -sf /repo/tests/config/executable_pool_dictionary.xml /etc/clickhouse-server/ + ln -sf /repo/tests/config/test_function.xml /etc/clickhouse-server/ + ln -sf /repo/tests/config/top_level_domains /etc/clickhouse-server/ + ln -sf /repo/tests/config/regions_hierarchy.txt /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/regions_names_en.txt /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/regions_names_es.txt /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/ext-en.txt /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/ext-ru.txt /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/lem-en.bin /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/server.key /etc/clickhouse-server/ + ln -sf /repo/tests/config/server.crt /etc/clickhouse-server/ + ln -sf /repo/tests/config/dhparam.pem /etc/clickhouse-server/ + ln -sf --backup=simple --suffix=_original.xml /repo/tests/config/config.d/query_masking_rules.xml /etc/clickhouse-server/config.d/ + [[ -n '' ]] + rm -f /etc/clickhouse-server/config.d/zookeeper_fault_injection.xml + ln -sf /repo/tests/config/config.d/zookeeper.xml /etc/clickhouse-server/config.d/ + [[ -n '' ]] + rm -f /etc/clickhouse-server/config.d/cannot_allocate_thread_injection.xml + value=1 + sed --follow-symlinks -i 's|[01]|1|' /etc/clickhouse-server/config.d/keeper_port.xml + value=67047424 + sed --follow-symlinks -i 's|[[:digit:]]\+|67047424|' /etc/clickhouse-server/config.d/keeper_port.xml + value=43378688 + sed --follow-symlinks -i 's|[[:digit:]]\+|43378688|' /etc/clickhouse-server/config.d/keeper_port.xml + [[ -n '' ]] + [[ -n '' ]] + [[ '' == \1 ]] + [[ '' == \1 ]] + [[ 1 == \1 ]] + [[ 0 != \1 ]] + ln -sf /repo/tests/config/config.d/azure_storage_conf.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/storage_conf.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/storage_conf_02944.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/storage_conf_02963.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/config.d/storage_conf_02961.xml /etc/clickhouse-server/config.d/ + ln -sf /repo/tests/config/users.d/s3_cache.xml /etc/clickhouse-server/users.d/ + ln -sf /repo/tests/config/users.d/s3_cache_new.xml /etc/clickhouse-server/users.d/ + [[ 0 == \1 ]] + ln -sf /repo/tests/config/client_config.xml /etc/clickhouse-client/config.xml + /repo/tests/docker_scripts/setup_minio.sh stateless + azurite-blob --blobHost 0.0.0.0 --blobPort 10000 --silent --inMemoryPersistence + export MINIO_ROOT_USER=clickhouse + MINIO_ROOT_USER=clickhouse + export MINIO_ROOT_PASSWORD=clickhouse + MINIO_ROOT_PASSWORD=clickhouse + main stateless + local query_dir ++ check_arg stateless ++ local query_dir ++ '[' '!' 1 -eq 1 ']' ++ case "$1" in ++ query_dir=0_stateless ++ echo 0_stateless + query_dir=0_stateless + '[' '!' -f ./minio ']' + start_minio + mkdir -p ./minio_data + ./minio --version minio version RELEASE.2024-08-03T04-33-23Z (commit-id=6efb56851c40da88d1ca15112e2d686a4ecec6b3) Runtime: go1.22.5 linux/amd64 License: GNU AGPLv3 - https://www.gnu.org/licenses/agpl-3.0.html Copyright: 2015-2024 MinIO, Inc. + wait_for_it + local counter=0 + local max_counter=60 + local url=http://localhost:11111 + ./minio server --address :11111 ./minio_data + params=('--silent' '--verbose') + local params + curl --silent --verbose http://localhost:11111 + grep AccessDenied trying to connect to minio + [[ 0 == \6\0 ]] + echo 'trying to connect to minio' + sleep 1 (node:273) [DEP0040] DeprecationWarning: The `punycode` module is deprecated. Please use a userland alternative instead. (Use `node --trace-deprecation ...` to show where the warning was created) In-memory extent storage is enabled with a limit of 15667.21 MB (16428261376 bytes, 50% of total memory). Azurite Blob service is starting on 0.0.0.0:10000 Azurite Blob service successfully listens on http://0.0.0.0:10000 INFO: Formatting 1st pool, 1 set(s), 1 drives per set. INFO: WARNING: Host local has more than 0 drives of set. A host failure will result in data becoming unavailable. MinIO Object Storage Server Copyright: 2015-2025 MinIO, Inc. License: GNU AGPLv3 - https://www.gnu.org/licenses/agpl-3.0.html Version: RELEASE.2024-08-03T04-33-23Z (go1.22.5 linux/amd64) API: http://172.17.0.2:11111 http://127.0.0.1:11111 WebUI: http://172.17.0.2:42161 http://127.0.0.1:42161 Docs: https://min.io/docs/minio/linux/index.html + counter=1 + curl --silent --verbose http://localhost:11111 + grep AccessDenied AccessDeniedAccess Denied./18324FA5B7C0C8967dc7eb22d3288ec80374614e9088e31d3668a6922ead55932dd2a8e56373820f + lsof -i :11111 COMMAND PID USER FD TYPE DEVICE SIZE/OFF NODE NAME minio 296 root 8u IPv4 32994 0t0 TCP localhost:11111 (LISTEN) minio 296 root 9u IPv6 32995 0t0 TCP *:11111 (LISTEN) minio 296 root 10u IPv6 32996 0t0 TCP localhost:11111 (LISTEN) + sleep 5 + setup_minio stateless + local test_type=stateless + ./mc alias set clickminio http://localhost:11111 clickhouse clickhouse Added `clickminio` successfully. + ./mc admin user add clickminio test testtest Added user `test` successfully. + ./mc admin policy attach clickminio readwrite --user=test Attached Policies: [readwrite] To User: test + ./mc mb --ignore-existing clickminio/test Bucket created successfully `clickminio/test`. + '[' stateless = stateless ']' + ./mc anonymous set public clickminio/test Access permission for `clickminio/test` is set to `public` + upload_data 0_stateless /repo/tests/ + local query_dir=0_stateless + local test_path=/repo/tests/ + local data_path=/repo/tests//queries/0_stateless/data_minio + '[' -d /repo/tests//queries/0_stateless/data_minio ']' + ./mc cp --recursive /repo/tests//queries/0_stateless/data_minio/ clickminio/test/ `/repo/tests/queries/0_stateless/data_minio/02731.parquet` -> `clickminio/test/02731.parquet` `/repo/tests/queries/0_stateless/data_minio/02876.parquet` -> `clickminio/test/02876.parquet` `/repo/tests/queries/0_stateless/data_minio/02366_data.jsonl` -> `clickminio/test/02366_data.jsonl` `/repo/tests/queries/0_stateless/data_minio/02731.arrow` -> `clickminio/test/02731.arrow` `/repo/tests/queries/0_stateless/data_minio/03036_archive1.tar` -> `clickminio/test/03036_archive1.tar` `/repo/tests/queries/0_stateless/data_minio/03036_archive1.zip` -> `clickminio/test/03036_archive1.zip` `/repo/tests/queries/0_stateless/data_minio/03036_archive2.tar` -> `clickminio/test/03036_archive2.tar` `/repo/tests/queries/0_stateless/data_minio/03036_archive2.zip` -> `clickminio/test/03036_archive2.zip` `/repo/tests/queries/0_stateless/data_minio/03036_archive3.tar.gz` -> `clickminio/test/03036_archive3.tar.gz` `/repo/tests/queries/0_stateless/data_minio/03036_compressed_file_archive.zip` -> `clickminio/test/03036_compressed_file_archive.zip` `/repo/tests/queries/0_stateless/data_minio/03036_json_archive.zip` -> `clickminio/test/03036_json_archive.zip` `/repo/tests/queries/0_stateless/data_minio/a.tsv` -> `clickminio/test/a.tsv` `/repo/tests/queries/0_stateless/data_minio/b.tsv` -> `clickminio/test/b.tsv` `/repo/tests/queries/0_stateless/data_minio/c.tsv` -> `clickminio/test/c.tsv` `/repo/tests/queries/0_stateless/data_minio/hive_partitioning/column0=Elizabeth/column1=Gordon/sample.parquet` -> `clickminio/test/hive_partitioning/column0=Elizabeth/column1=Gordon/sample.parquet` `/repo/tests/queries/0_stateless/data_minio/hive_partitioning/column0=Elizabeth/column1=Schmidt/sample.parquet` -> `clickminio/test/hive_partitioning/column0=Elizabeth/column1=Schmidt/sample.parquet` `/repo/tests/queries/0_stateless/data_minio/hive_partitioning/column0=Elizabeth/sample.parquet` -> `clickminio/test/hive_partitioning/column0=Elizabeth/sample.parquet` `/repo/tests/queries/0_stateless/data_minio/hive_partitioning/non_existing_column=Elizabeth/sample.parquet` -> `clickminio/test/hive_partitioning/non_existing_column=Elizabeth/sample.parquet` `/repo/tests/queries/0_stateless/data_minio/json_data` -> `clickminio/test/json_data` `/repo/tests/queries/0_stateless/data_minio/tsv_with_header.tsv` -> `clickminio/test/tsv_with_header.tsv` Total: 5.42 MiB, Transferred: 5.42 MiB, Speed: 128.74 MiB/s + setup_aws_credentials + local minio_root_user=clickhouse + local minio_root_password=clickhouse + mkdir -p /root/.aws + cat + config_logs_export_cluster /etc/clickhouse-server/config.d/system_logs_export.yaml + set +x File /tmp/export-logs-config.sh does not exist, do not setup + [[ -n '' ]] + export IS_FLAKY_CHECK=0 + IS_FLAKY_CHECK=0 + export NUM_TRIES + '[' 1 -gt 1 ']' + sudo -E -u clickhouse /usr/bin/clickhouse-server --config /etc/clickhouse-server/config.xml --daemon --pid-file /var/run/clickhouse-server/clickhouse-server.pid + [[ 0 -eq 1 ]] + [[ 0 -eq 1 ]] + for _ in {1..100} + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + sleep 1 + for _ in {1..100} + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + sleep 1 + for _ in {1..100} + clickhouse-client --query 'SELECT 1' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + sleep 1 + for _ in {1..100} + clickhouse-client --query 'SELECT 1' 1 + break + setup_logs_replication + set +x File /tmp/export-logs-config.sh does not exist, do not setup + attach_gdb_to_clickhouse ++ run_with_retry 5 clickhouse-client --query 'SELECT count() FROM system.build_options WHERE name = '\''CXX_FLAGS'\'' AND position('\''sanitize=address'\'' IN value)' ++ [[ ahxB =~ e ]] ++ set_e=false ++ set +e ++ local total_retries=5 ++ shift ++ local retry=0 ++ '[' 0 -ge 5 ']' ++ clickhouse-client --query 'SELECT count() FROM system.build_options WHERE name = '\''CXX_FLAGS'\'' AND position('\''sanitize=address'\'' IN value)' ++ false ++ return + IS_ASAN=0 + [[ 0 = \1 ]] ++ kill -l SIGRTMIN + RTMIN=34 + echo ' set follow-fork-mode parent handle SIGHUP nostop noprint pass handle SIGINT nostop noprint pass handle SIGQUIT nostop noprint pass handle SIGPIPE nostop noprint pass handle SIGTERM nostop noprint pass handle SIGUSR1 nostop noprint pass handle SIGUSR2 nostop noprint pass handle SIG34 nostop noprint pass info signals continue backtrace full info registers p top' 1 KiB of the 'stack: p/x *(uint64_t[128]*)$sp maintenance info sections thread apply all backtrace full disassemble /s up disassemble /s up disassemble /s p "done" detach quit ' + sleep 5 + ts '%Y-%m-%d %H:%M:%S' ++ cat /var/run/clickhouse-server/clickhouse-server.pid + gdb -batch -command script.gdb -p 428 aarch64-binfmt-P: Could not open '/lib/ld-linux-aarch64.so.1': No such file or directory + run_with_retry 60 clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' + [[ aehxB =~ e ]] + set_e=true + set +e + local total_retries=60 + shift + local retry=0 + '[' 0 -ge 60 ']' + clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\''' Connected to clickhouse-server after attaching gdb + true + set -e + return + clickhouse-client --allow_experimental_json_type=1 --query 'CREATE TABLE minio_audit_logs ( log JSON(time DateTime64(9)) ) ENGINE = MergeTree ORDER BY tuple()' + clickhouse-client --allow_experimental_json_type=1 --query 'CREATE TABLE minio_server_logs ( log JSON(time DateTime64(9)) ) ENGINE = MergeTree ORDER BY tuple()' + ./mc admin config set clickminio logger_webhook:ch_server_webhook 'endpoint=http://localhost:8123/?async_insert=1&wait_for_async_insert=0&async_insert_busy_timeout_min_ms=5000&async_insert_busy_timeout_max_ms=5000&async_insert_max_query_number=1000&async_insert_max_data_size=10485760&date_time_input_format=best_effort&query=INSERT%20INTO%20minio_server_logs%20FORMAT%20JSONAsObject' queue_size=1000000 batch_size=500 Successfully applied new settings. + ./mc admin config set clickminio audit_webhook:ch_audit_webhook 'endpoint=http://localhost:8123/?async_insert=1&wait_for_async_insert=0&async_insert_busy_timeout_min_ms=5000&async_insert_busy_timeout_max_ms=5000&async_insert_max_query_number=1000&async_insert_max_data_size=10485760&date_time_input_format=best_effort&query=INSERT%20INTO%20minio_audit_logs%20FORMAT%20JSONAsObject' queue_size=1000000 batch_size=500 Successfully applied new settings. + max_retries=100 + retry=1 + '[' 1 -le 100 ']' + echo 'clickminio restart attempt 1:' clickminio restart attempt 1: ++ ./mc admin service restart clickminio --wait --json ++ jq -r .status INFO: Restarting on service signal MinIO Object Storage Server Copyright: 2015-2025 MinIO, Inc. License: GNU AGPLv3 - https://www.gnu.org/licenses/agpl-3.0.html Version: RELEASE.2024-08-03T04-33-23Z (go1.22.5 linux/amd64) API: http://172.17.0.2:11111 http://127.0.0.1:11111 WebUI: http://172.17.0.2:36665 http://127.0.0.1:36665 Docs: https://min.io/docs/minio/linux/index.html Output of restart status: success success Restarted clickminio successfully. + output='success success' + echo 'Output of restart status: success success' + expected_output='success success' + '[' 'success success' = 'success success' ']' + echo 'Restarted clickminio successfully.' + break + '[' 1 -gt 100 ']' + MC_ADMIN_PID=1344 + ./mc admin trace clickminio + export -f run_tests + '[' 1 -gt 1 ']' + run_tests + set -x + read -ra ADDITIONAL_OPTIONS + HIGH_LEVEL_COVERAGE=YES + '[' 1 -gt 1 ']' + [[ -n '' ]] + [[ -n '' ]] + [[ 0 -eq 1 ]] + [[ '' -eq 1 ]] + [[ 0 -eq 1 ]] ++ clickhouse-client --query 'SELECT value LIKE '\''%SANITIZE_COVERAGE%'\'' FROM system.build_options WHERE name = '\''CXX_FLAGS'\''' + [[ 1 == 0 ]] + ADDITIONAL_OPTIONS+=('--jobs') + ADDITIONAL_OPTIONS+=('8') + [[ -n 2 ]] + [[ -n 4 ]] + ADDITIONAL_OPTIONS+=('--run-by-hash-num') + ADDITIONAL_OPTIONS+=("$RUN_BY_HASH_NUM") + ADDITIONAL_OPTIONS+=('--run-by-hash-total') + ADDITIONAL_OPTIONS+=("$RUN_BY_HASH_TOTAL") + HIGH_LEVEL_COVERAGE=NO + [[ -n '' ]] + [[ NO = \Y\E\S ]] + ADDITIONAL_OPTIONS+=('--report-logs-stats') + try_run_with_retry 10 clickhouse-client -q 'insert into system.zookeeper (name, path, value) values ('\''auxiliary_zookeeper2'\'', '\''/test/chroot/'\'', '\'''\'')' + local total_retries=10 + shift + fn_exists run_with_retry + declare -F run_with_retry + run_with_retry 10 clickhouse-client -q 'insert into system.zookeeper (name, path, value) values ('\''auxiliary_zookeeper2'\'', '\''/test/chroot/'\'', '\'''\'')' + [[ aehxB =~ e ]] + set_e=true + set +e + local total_retries=10 + shift + local retry=0 + '[' 0 -ge 10 ']' + clickhouse-client -q 'insert into system.zookeeper (name, path, value) values ('\''auxiliary_zookeeper2'\'', '\''/test/chroot/'\'', '\'''\'')' + true + set -e + return + set +e + TEST_ARGS=(--testname --shard --zookeeper --check-zookeeper-session --hung-check --print-time --no-drop-if-fail --capture-client-stacktrace --queries "/repo/tests/queries" --test-runs "$NUM_TRIES" "${ADDITIONAL_OPTIONS[@]}") + clickhouse-test --testname --shard --zookeeper --check-zookeeper-session --hung-check --print-time --no-drop-if-fail --capture-client-stacktrace --queries /repo/tests/queries --test-runs 1 --hung-check --print-time --jobs 8 --run-by-hash-num 2 --run-by-hash-total 4 --report-logs-stats + ts '%Y-%m-%d %H:%M:%S' + tee -a test_output/test_result.txt 2025-04-01 17:39:56 Using queries from '/repo/tests/queries' directory 2025-04-01 17:39:56 Connecting to ClickHouse server... OK 2025-04-01 17:39:56 Connected to server 24.12.2.20221.altinityantalya @ 82252d159dc02cab0f366aaa5691adc1545dd11d HEAD 2025-04-01 17:39:56 Found 1667 parallel tests and 130 sequential tests 2025-04-01 17:39:57 Running about 208 stateless tests (Process-8). 2025-04-01 17:39:57 03085_analyzer_alias_column_group_by: [ OK ] 0.58 sec. 2025-04-01 17:39:57 Running about 208 stateless tests (Process-10). 2025-04-01 17:39:57 01906_partition_by_multiply_by_zero: [ OK ] 0.73 sec. 2025-04-01 17:39:57 Running about 208 stateless tests (Process-6). 2025-04-01 17:39:57 03171_direct_dict_short_circuit_bug: [ OK ] 0.83 sec. 2025-04-01 17:39:57 01851_hedged_connections_external_tables: [ SKIPPED ] 0.00 sec. 2025-04-01 17:39:57 Reason: not running for current build 2025-04-01 17:39:57 Running about 208 stateless tests (Process-5). 2025-04-01 17:39:57 01010_pm_join_all_join_bug: [ OK ] 0.89 sec. 2025-04-01 17:39:57 Running about 208 stateless tests (Process-9). 2025-04-01 17:39:57 02724_persist_interval_type: [ OK ] 0.88 sec. 2025-04-01 17:39:57 Running about 208 stateless tests (Process-3). 2025-04-01 17:39:57 00228_shard_quantiles_deterministic_merge_overflow: [ OK ] 1.44 sec. 2025-04-01 17:39:58 02815_first_line: [ OK ] 0.68 sec. 2025-04-01 17:39:58 01549_low_cardinality_mv_fuzz: [ OK ] 0.68 sec. 2025-04-01 17:39:58 00251_has_types: [ OK ] 0.78 sec. 2025-04-01 17:39:58 02372_now_in_block: [ OK ] 1.18 sec. 2025-04-01 17:39:58 00820_multiple_joins_subquery_requires_alias: [ OK ] 1.13 sec. 2025-04-01 17:39:58 01914_ubsan_quantile_timing: [ OK ] 0.53 sec. 2025-04-01 17:39:58 02815_alias_to_length: [ OK ] 0.53 sec. 2025-04-01 17:39:58 00492_drop_temporary_table: [ OK ] 0.62 sec. 2025-04-01 17:39:59 00361_shared_array_offsets_and_squash_blocks: [ OK ] 0.73 sec. 2025-04-01 17:39:59 03230_subcolumns_mv: [ OK ] 0.73 sec. 2025-04-01 17:39:59 02465_limit_trivial_max_rows_to_read: [ OK ] 0.88 sec. 2025-04-01 17:40:00 Running about 208 stateless tests (Process-4). 2025-04-01 17:40:00 03230_keeper_cp_mv_commands: [ OK ] 3.49 sec. 2025-04-01 17:40:00 02723_param_exception_message_context: [ OK ] 2.13 sec. 2025-04-01 17:40:01 01043_h3_edge_length_m: [ OK ] 0.58 sec. 2025-04-01 17:40:01 00545_weird_aggregate_functions: [ OK ] 0.58 sec. 2025-04-01 17:40:02 01196_max_parser_depth: [ OK ] 2.63 sec. 2025-04-01 17:40:02 02352_interactive_queries_from_file: [ OK ] 2.13 sec. 2025-04-01 17:40:02 01436_storage_merge_with_join_push_down: [ OK ] 0.83 sec. 2025-04-01 17:40:02 Running about 208 stateless tests (Process-7). 2025-04-01 17:40:02 03031_minmax_index_for_pointinpolygon: [ OK ] 6.29 sec. 2025-04-01 17:40:02 00276_sample: [ OK ] 4.48 sec. 2025-04-01 17:40:03 02012_sha512_fixedstring: [ OK ] 0.63 sec. 2025-04-01 17:40:03 01036_union_different_columns: [ OK ] 0.68 sec. 2025-04-01 17:40:04 03285_analyzer_extract_common_expr_bug: [ OK ] 0.63 sec. 2025-04-01 17:40:04 02250_lots_of_columns_in_csv_with_names: [ OK ] 4.99 sec. 2025-04-01 17:40:05 03237_get_subcolumn_low_cardinality_column: [ OK ] 0.53 sec. 2025-04-01 17:40:05 02157_readonly_system_suspend: [ OK ] 1.98 sec. 2025-04-01 17:40:06 00122_join_with_subquery_with_subquery: [ OK ] 0.58 sec. 2025-04-01 17:40:06 01505_trivial_count_with_partition_predicate: [ OK ] 1.18 sec. 2025-04-01 17:40:06 03262_analyzer_materialized_view_in_with_cte: [ OK ] 0.73 sec. 2025-04-01 17:40:06 03198_unload_primary_key_outdated: [ OK ] 4.69 sec. 2025-04-01 17:40:07 00396_uuid: [ OK ] 0.73 sec. 2025-04-01 17:40:08 01949_heredoc_unfinished: [ OK ] 1.64 sec. 2025-04-01 17:40:08 01601_proxy_protocol: [ OK ] 1.48 sec. 2025-04-01 17:40:09 00365_statistics_in_formats: [ OK ] 5.04 sec. 2025-04-01 17:40:09 01214_test_storage_merge_aliases_with_where: [ OK ] 1.13 sec. 2025-04-01 17:40:10 02366_window_function_order_by: [ OK ] 0.68 sec. 2025-04-01 17:40:10 00271_agg_state_and_totals: [ OK ] 0.68 sec. 2025-04-01 17:40:10 03096_largest_triangle_3b_crash: [ OK ] 0.53 sec. 2025-04-01 17:40:11 03064_analyzer_named_subqueries: [ OK ] 0.58 sec. 2025-04-01 17:40:11 02158_proportions_ztest_cmp: [ OK ] 2.68 sec. 2025-04-01 17:40:11 01866_datetime64_cmp_with_constant: [ OK ] 1.08 sec. 2025-04-01 17:40:11 02843_backup_use_same_password_for_base_backup: [ OK ] 9.55 sec. 2025-04-01 17:40:11 03263_analyzer_materialized_view_cte_nested: [ OK ] 0.73 sec. 2025-04-01 17:40:13 02122_parallel_formatting_JSONCompactEachRowWithNames: [ OK ] 6.75 sec. 2025-04-01 17:40:14 02293_grouping_function_group_by: [ OK ] 2.54 sec. 2025-04-01 17:40:14 01851_clear_column_referenced_by_mv: [ OK ] 0.78 sec. 2025-04-01 17:40:14 00652_replicated_mutations_default_database_zookeeper: [ OK ] 3.69 sec. 2025-04-01 17:40:15 01416_join_totals_header_bug: [ OK ] 0.73 sec. 2025-04-01 17:40:15 01825_type_json_18: [ OK ] 0.64 sec. 2025-04-01 17:40:16 02538_alter_rename_sequence: [ OK ] 1.43 sec. 2025-04-01 17:40:18 01018_Distributed__shard_num: [ OK ] 2.39 sec. 2025-04-01 17:40:18 01289_min_execution_speed_not_too_early: [ OK ] 15.47 sec. 2025-04-01 17:40:18 00823_sequence_match_dfa: [ OK ] 6.40 sec. 2025-04-01 17:40:19 02515_and_or_if_multiif_not_return_lc: [ OK ] 0.57 sec. 2025-04-01 17:40:19 01769_extended_range_2: [ OK ] 0.83 sec. 2025-04-01 17:40:19 00050_any_left_join: [ OK ] 0.58 sec. 2025-04-01 17:40:19 03165_storage_merge_view_prewhere: [ OK ] 0.78 sec. 2025-04-01 17:40:20 03068_analyzer_distributed_join: [ OK ] 0.88 sec. 2025-04-01 17:40:21 02433_default_expression_operator_in: [ OK ] 1.08 sec. 2025-04-01 17:40:22 02861_alter_replace_partition_do_not_wait_mutations_on_unrelated_partitions: [ OK ] 7.09 sec. 2025-04-01 17:40:22 02932_group_by_null_fuzzer: [ OK ] 0.63 sec. 2025-04-01 17:40:22 00369_int_div_of_float: [ OK ] 0.58 sec. 2025-04-01 17:40:22 01455_nullable_type_with_if_agg_combinator: [ OK ] 0.58 sec. 2025-04-01 17:40:23 02916_replication_protocol_wait_for_part: [ OK ] 11.20 sec. 2025-04-01 17:40:23 01762_deltasumtimestamp: [ OK ] 0.73 sec. 2025-04-01 17:40:24 03033_dist_settings.optimize_skip_unused_shards_rewrite_in_composite_sharding_key: [ OK ] 0.78 sec. 2025-04-01 17:40:24 01273_arrow_nested_arrays_load: [ OK ] 4.58 sec. 2025-04-01 17:40:25 01303_polygons_equals: [ OK ] 0.63 sec. 2025-04-01 17:40:26 00969_columns_clause: [ OK ] 1.38 sec. 2025-04-01 17:40:26 00752_low_cardinality_permute: [ OK ] 0.89 sec. 2025-04-01 17:40:26 02516_projections_with_rollup: [ OK ] 28.20 sec. 2025-04-01 17:40:26 01921_with_fill_with_totals: [ OK ] 0.68 sec. 2025-04-01 17:40:27 02158_ztest: [ OK ] 0.70 sec. 2025-04-01 17:40:27 00900_orc_arrow_parquet_tuples: [ OK ] 9.30 sec. 2025-04-01 17:40:27 02723_jit_aggregation_bug_48120: [ OK ] 0.93 sec. 2025-04-01 17:40:28 00619_union_highlite: [ OK ] 0.79 sec. 2025-04-01 17:40:29 02534_parquet_fixed_binary_array: [ OK ] 6.19 sec. 2025-04-01 17:40:30 01045_array_zip: [ OK ] 1.24 sec. 2025-04-01 17:40:30 01477_lc_in_merge_join_left_key: [ OK ] 2.74 sec. 2025-04-01 17:40:31 01413_alter_update_supertype: [ OK ] 1.29 sec. 2025-04-01 17:40:32 02813_seriesDecomposeSTL: [ OK ] 1.75 sec. 2025-04-01 17:40:32 02024_compression_in_query: [ OK ] 6.82 sec. 2025-04-01 17:40:32 02985_if_over_big_int_decimal: [ OK ] 0.94 sec. 2025-04-01 17:40:33 00942_mv_rename_table: [ OK ] 0.84 sec. 2025-04-01 17:40:33 02113_untuple_func_alias: [ OK ] 0.59 sec. 2025-04-01 17:40:34 02030_function_mapContainsKeyLike: [ OK ] 1.12 sec. 2025-04-01 17:40:34 00007_array: [ OK ] 0.59 sec. 2025-04-01 17:40:35 02212_h3_get_pentagon_indexes: [ OK ] 1.09 sec. 2025-04-01 17:40:35 02129_add_column_add_ttl: [ OK ] 2.17 sec. 2025-04-01 17:40:36 03034_recursive_cte_tree: [ OK ] 0.89 sec. 2025-04-01 17:40:36 00723_remerge_sort: [ OK ] 8.70 sec. 2025-04-01 17:40:37 02733_sparse_columns_reload: [ OK ] 1.04 sec. 2025-04-01 17:40:37 00380_client_break_at_exception_in_batch_mode: [ OK ] 2.14 sec. 2025-04-01 17:40:38 02496_remove_redundant_sorting_analyzer: [ OK ] 22.09 sec. 2025-04-01 17:40:40 02952_clickhouse_local_query_parameters_cli: [ OK ] 1.98 sec. 2025-04-01 17:40:41 00700_to_decimal_or_something_1: [ OK ] 4.50 sec. 2025-04-01 17:40:42 02751_text_formats_bad_nullable_parsing: [ OK ] 5.55 sec. 2025-04-01 17:40:43 02135_local_create_db: [ OK ] 2.24 sec. 2025-04-01 17:40:44 02122_4letter_words_stress_zookeeper: [ OK ] 21.44 sec. 2025-04-01 17:40:44 01551_mergetree_read_in_order_spread: [ OK ] 0.88 sec. 2025-04-01 17:40:45 02870_per_column_settings: [ OK ] 1.23 sec. 2025-04-01 17:40:46 02316_hierarchical_dictionaries_nullable_parent_key: [ OK ] 1.93 sec. 2025-04-01 17:40:47 01062_pm_all_join_with_block_continuation: [ OK ] 48.31 sec. 2025-04-01 17:40:49 02498_random_string_in_json_schema_inference: [ OK ] 1.93 sec. 2025-04-01 17:40:52 02149_schema_inference_create_table_syntax: [ OK ] 9.80 sec. 2025-04-01 17:40:53 01499_log_deadlock: [ OK ] 0.74 sec. 2025-04-01 17:40:54 01655_plan_optimizations: [ OK ] 26.02 sec. 2025-04-01 17:40:54 00666_uniq_complex_types: [ OK ] 1.18 sec. 2025-04-01 17:40:55 02999_scalar_subqueries_bug_2: [ OK ] 0.58 sec. 2025-04-01 17:40:56 02540_date_column_consistent_insert_behaviour: [ OK ] 1.88 sec. 2025-04-01 17:40:56 02864_statistics_predicates: [ OK ] 11.27 sec. 2025-04-01 17:40:57 02295_GROUP_BY_AggregateFunction: [ OK ] 0.94 sec. 2025-04-01 17:40:57 00157_aliases_and_lambda_formal_parameters: [ OK ] 0.48 sec. 2025-04-01 17:40:58 01938_joins_identifiers: [ OK ] 0.73 sec. 2025-04-01 17:40:58 01505_pipeline_executor_UAF: [ OK ] 23.82 sec. 2025-04-01 17:40:59 02311_range_hashed_dictionary_range_cast: [ OK ] 0.69 sec. 2025-04-01 17:40:59 03036_reading_s3_archives: [ OK ] 2.39 sec. 2025-04-01 17:40:59 03046_column_in_block_array_join: [ OK ] 0.68 sec. 2025-04-01 17:40:59 00647_histogram: [ OK ] 0.68 sec. 2025-04-01 17:40:59 00581_limit_on_result_and_subquery_and_insert: [ OK ] 0.63 sec. 2025-04-01 17:40:59 02766_bitshift_with_const_arguments: [ OK ] 0.83 sec. 2025-04-01 17:41:00 03205_system_sync_replica_format: [ OK ] 0.58 sec. 2025-04-01 17:41:00 03071_analyzer_array_join_forbid_non_existing_columns: [ OK ] 0.79 sec. 2025-04-01 17:41:01 03162_dynamic_type_nested: [ OK ] 0.63 sec. 2025-04-01 17:41:01 01458_named_tuple_millin: [ OK ] 0.53 sec. 2025-04-01 17:41:01 03257_async_insert_native_empty_block: [ OK ] 6.69 sec. 2025-04-01 17:41:01 03224_json_merges_new_type_in_shared_data: [ OK ] 0.78 sec. 2025-04-01 17:41:01 03270_processors_profile_log_3: [ OK ] 12.41 sec. 2025-04-01 17:41:02 01010_partial_merge_join_const_and_lc: [ OK ] 0.78 sec. 2025-04-01 17:41:04 01780_column_sparse_full: [ OK ] 2.68 sec. 2025-04-01 17:41:04 01339_client_unrecognized_option: [ OK ] 2.34 sec. 2025-04-01 17:41:04 02020_cast_integer_overflow: [ OK ] 0.53 sec. 2025-04-01 17:41:04 00003_reinterpret_as_string: [ OK ] 0.53 sec. 2025-04-01 17:41:05 02294_dictionaries_hierarchical_index: [ OK ] 1.08 sec. 2025-04-01 17:41:06 01214_point_in_Mecca: [ OK ] 4.24 sec. 2025-04-01 17:41:06 02499_analyzer_set_index: [ OK ] 0.68 sec. 2025-04-01 17:41:07 01926_date_date_time_supertype: [ OK ] 0.68 sec. 2025-04-01 17:41:08 01825_type_json_missed_values: [ OK ] 1.18 sec. 2025-04-01 17:41:10 00339_parsing_bad_arrays: [ OK ] 1.48 sec. 2025-04-01 17:41:10 01098_msgpack_format: [ OK ] 29.96 sec. 2025-04-01 17:41:10 02896_leading_zeroes_no_octal: [ OK ] 4.33 sec. 2025-04-01 17:41:11 02513_prewhere_combine_step_filters: [ OK ] 1.13 sec. 2025-04-01 17:41:11 02508_index_analysis_to_date_timezone: [ OK ] 0.83 sec. 2025-04-01 17:41:11 00030_alter_table: [ OK ] 1.48 sec. 2025-04-01 17:41:12 02788_current_schemas_function: [ OK ] 0.88 sec. 2025-04-01 17:41:12 03024_total_rows_approx_is_set_for_system_zeros_and_generate_random: [ OK ] 0.73 sec. 2025-04-01 17:41:12 02864_statistics_delayed_materialization_in_merge: [ OK ] 1.08 sec. 2025-04-01 17:41:13 03243_array_join_lambda: [ OK ] 0.58 sec. 2025-04-01 17:41:13 03203_variant_convert_field_to_type_bug: [ OK ] 0.63 sec. 2025-04-01 17:41:13 03239_if_constant_folding: [ OK ] 0.74 sec. 2025-04-01 17:41:13 02534_join_prewhere_bug: [ OK ] 0.88 sec. 2025-04-01 17:41:14 01846_alter_column_without_type_bugfix: [ OK ] 0.68 sec. 2025-04-01 17:41:14 00552_logical_functions_uint8_as_bool: [ OK ] 0.68 sec. 2025-04-01 17:41:15 02324_map_combinator_bug: [ OK ] 1.13 sec. 2025-04-01 17:41:16 01710_projection_with_column_transformers: [ OK ] 0.53 sec. 2025-04-01 17:41:18 02572_query_views_log_background_thread: [ OK ] 18.18 sec. 2025-04-01 17:41:19 02789_functions_after_sorting_and_columns_with_same_names_bug: [ OK ] 1.03 sec. 2025-04-01 17:41:19 03230_async_insert_native: [ OK ] 4.89 sec. 2025-04-01 17:41:19 02472_segfault_expression_parser: [ OK ] 0.43 sec. 2025-04-01 17:41:20 02481_i43247_ubsan_in_minmaxany: [ OK ] 1.39 sec. 2025-04-01 17:41:21 02718_parquet_metadata_format: [ OK ] 5.01 sec. 2025-04-01 17:41:22 01188_attach_table_from_path: [ OK ] 1.13 sec. 2025-04-01 17:41:23 02715_or_null: [ OK ] 0.58 sec. 2025-04-01 17:41:23 02269_to_start_of_interval_overflow: [ OK ] 0.68 sec. 2025-04-01 17:41:24 02806_cte_block_cannot_be_empty: [ OK ] 0.64 sec. 2025-04-01 17:41:25 02246_flatten_tuple: [ OK ] 0.93 sec. 2025-04-01 17:41:30 02439_merge_selecting_partitions: [ OK ] 10.11 sec. 2025-04-01 17:41:36 01338_long_select_and_alter_zookeeper: [ OK ] 16.06 sec. 2025-04-01 17:41:38 00661_optimize_final_replicated_without_partition_zookeeper: [ OK ] 1.24 sec. 2025-04-01 17:41:38 02029_quantile_sanitizer: [ OK ] 0.73 sec. 2025-04-01 17:41:39 00552_or_nullable: [ OK ] 0.88 sec. 2025-04-01 17:41:41 03271_ghdata_object_to_json_alter: [ OK ] 63.87 sec. 2025-04-01 17:41:41 03174_least_greatest_ignore_null_input_values: [ OK ] 1.33 sec. 2025-04-01 17:41:41 02243_in_ip_address: [ OK ] 0.64 sec. 2025-04-01 17:41:42 03148_setting_max_streams_to_max_threads_ratio_overflow: [ OK ] 1.08 sec. 2025-04-01 17:41:45 01581_deduplicate_by_columns_local: [ OK ] 2.89 sec. 2025-04-01 17:41:45 02346_read_in_order_fixed_prefix: [ OK ] 40.92 sec. 2025-04-01 17:41:45 03271_decimal_monotonic_day_of_week: [ OK ] 0.63 sec. 2025-04-01 17:41:47 02574_suspicious_low_cardinality_msan: [ OK ] 1.28 sec. 2025-04-01 17:41:47 03268_system_parts_index_granularity: [ OK ] 0.54 sec. 2025-04-01 17:41:47 02956_clickhouse_local_system_parts: [ OK ] 2.23 sec. 2025-04-01 17:41:48 01710_query_log_with_projection_info: [ OK ] 6.20 sec. 2025-04-01 17:41:48 01802_toDateTime64_large_values: [ OK ] 0.68 sec. 2025-04-01 17:41:48 00700_decimal_in_keys: [ OK ] 1.08 sec. 2025-04-01 17:41:50 02791_final_block_structure_mismatch_bug: [ OK ] 1.88 sec. 2025-04-01 17:41:50 00411_long_accurate_number_comparison_int1: [ OK ] 64.04 sec. 2025-04-01 17:41:50 02680_default_star: [ OK ] 0.43 sec. 2025-04-01 17:41:50 01825_type_json_ghdata_insert_select: [ OK ] 47.89 sec. 2025-04-01 17:41:50 01423_if_nullable_cond: [ OK ] 0.54 sec. 2025-04-01 17:41:51 01544_file_engine_settings: [ OK ] 2.48 sec. 2025-04-01 17:41:51 02833_std_alias: [ OK ] 0.63 sec. 2025-04-01 17:41:51 02204_fractional_progress_bar_long: [ SKIPPED ] 0.00 sec. 2025-04-01 17:41:51 Reason: not running for current build 2025-04-01 17:41:51 00004_shard_format_ast_and_remote_table: [ OK ] 0.64 sec. 2025-04-01 17:41:51 00804_test_custom_compression_codes_log_storages: [ OK ] 2.33 sec. 2025-04-01 17:41:51 00730_unicode_terminal_format: [ OK ] 0.83 sec. 2025-04-01 17:41:52 02245_s3_virtual_columns: [ OK ] 0.93 sec. 2025-04-01 17:41:52 02566_analyzer_limit_settings_distributed: [ OK ] 0.98 sec. 2025-04-01 17:41:52 03237_max_map_state_decimal_serialization: [ OK ] 0.58 sec. 2025-04-01 17:41:52 00353_join_by_tuple: [ OK ] 0.53 sec. 2025-04-01 17:41:52 01354_tuple_low_cardinality_array_mapped_bug: [ OK ] 0.58 sec. 2025-04-01 17:41:53 02922_server_exit_code: [ OK ] 2.13 sec. 2025-04-01 17:41:54 02417_null_variadic_behaviour: [ OK ] 1.98 sec. 2025-04-01 17:41:55 01834_alias_columns_laziness_filimonov: [ OK ] 3.10 sec. 2025-04-01 17:41:56 01731_async_task_queue_wait: [ OK ] 3.24 sec. 2025-04-01 17:41:56 02097_polygon_dictionary_store_key: [ OK ] 0.98 sec. 2025-04-01 17:41:57 00002_system_numbers: [ OK ] 0.83 sec. 2025-04-01 17:41:57 03008_deduplication_insert_into_partitioned_table: [ OK ] 2.54 sec. 2025-04-01 17:41:57 02982_parallel_replicas_unexpected_cluster: [ OK ] 0.73 sec. 2025-04-01 17:41:58 01070_alter_with_ttl: [ OK ] 0.68 sec. 2025-04-01 17:41:58 02351_Map_combinator_dist: [ OK ] 1.29 sec. 2025-04-01 17:41:59 00520_http_nullable: [ OK ] 1.59 sec. 2025-04-01 17:41:59 00908_bloom_filter_index: [ OK ] 29.45 sec. 2025-04-01 17:41:59 03206_no_exceptions_clickhouse_local: [ FAIL ] 1.78 sec. 2025-04-01 17:41:59 Reason: return code: 134, result: 2025-04-01 17:41:59 2025-04-01 17:41:59 2025-04-01 17:41:59 2025-04-01 17:41:59 stdout: 2025-04-01 17:41:59 2025-04-01 17:41:59 2025-04-01 17:41:59 Settings used in the test: --max_insert_threads 2 --group_by_two_level_threshold 672595 --group_by_two_level_threshold_bytes 22049239 --distributed_aggregation_memory_efficient 1 --fsync_metadata 0 --output_format_parallel_formatting 1 --input_format_parallel_parsing 0 --min_chunk_bytes_for_parallel_parsing 10871429 --max_read_buffer_size 994986 --prefer_localhost_replica 0 --max_block_size 90665 --max_joined_block_size_rows 93016 --max_threads 3 --optimize_append_index 1 --optimize_if_chain_to_multiif 0 --optimize_if_transform_strings_to_enum 1 --optimize_read_in_order 1 --optimize_or_like_chain 1 --optimize_substitute_columns 1 --enable_multiple_prewhere_read_steps 1 --read_in_order_two_level_merge_threshold 21 --optimize_aggregation_in_order 1 --aggregation_in_order_max_block_bytes 10920486 --use_uncompressed_cache 0 --min_bytes_to_use_direct_io 10737418240 --min_bytes_to_use_mmap_io 10737418240 --local_filesystem_read_method mmap --remote_filesystem_read_method read --local_filesystem_read_prefetch 0 --filesystem_cache_segments_batch_size 0 --read_from_filesystem_cache_if_exists_otherwise_bypass_cache 1 --throw_on_error_from_cache_on_write_operations 0 --remote_filesystem_read_prefetch 1 --allow_prefetched_read_pool_for_remote_filesystem 0 --filesystem_prefetch_max_memory_usage 32Mi --filesystem_prefetches_limit 0 --filesystem_prefetch_min_bytes_for_single_read_task 16Mi --filesystem_prefetch_step_marks 0 --filesystem_prefetch_step_bytes 0 --compile_aggregate_expressions 0 --compile_sort_description 1 --merge_tree_coarse_index_granularity 19 --optimize_distinct_in_order 1 --max_bytes_before_external_sort 0 --max_bytes_before_external_group_by 10737418240 --max_bytes_before_remerge_sort 787984612 --min_compress_block_size 1636648 --max_compress_block_size 2024248 --merge_tree_compact_parts_min_granules_to_multibuffer_read 13 --optimize_sorting_by_input_stream_properties 0 --http_response_buffer_size 8871796 --http_wait_end_of_query False --enable_memory_bound_merging_of_aggregation_results 1 --min_count_to_compile_expression 3 --min_count_to_compile_aggregate_expression 3 --min_count_to_compile_sort_description 3 --session_timezone Africa/Khartoum --use_page_cache_for_disks_without_file_cache False --page_cache_inject_eviction False --merge_tree_read_split_ranges_into_intersecting_and_non_intersecting_injection_probability 0.36 --prefer_external_sort_block_bytes 100000000 --cross_join_min_rows_to_compress 100000000 --cross_join_min_bytes_to_compress 0 --min_external_table_block_size_bytes 1 --max_parsing_threads 10 --optimize_functions_to_subcolumns 0 --parallel_replicas_local_plan 1 --query_plan_join_swap_table false --output_format_native_write_json_as_string 1 --enable_vertical_final 0 2025-04-01 17:41:59 2025-04-01 17:41:59 MergeTree settings used in test: --ratio_of_defaults_for_sparse_serialization 1.0 --prefer_fetch_merged_part_size_threshold 2088974808 --vertical_merge_algorithm_min_rows_to_activate 1000000 --vertical_merge_algorithm_min_columns_to_activate 1 --allow_vertical_merges_from_compact_to_wide_parts 0 --min_merge_bytes_to_use_direct_io 531140840 --index_granularity_bytes 7269453 --merge_max_block_size 6689 --index_granularity 58306 --min_bytes_for_wide_part 0 --marks_compress_block_size 45718 --primary_key_compress_block_size 81099 --replace_long_file_name_to_hash 0 --max_file_name_length 24 --min_bytes_for_full_part_storage 536870912 --compact_parts_max_bytes_to_buffer 35988816 --compact_parts_max_granules_to_buffer 120 --compact_parts_merge_max_bytes_to_prefetch_part 27544159 --cache_populated_by_fetch 0 --concurrent_part_removal_threshold 0 --old_parts_lifetime 10 --prewarm_mark_cache 1 --use_const_adaptive_granularity 0 --enable_index_granularity_compression 1 --use_primary_key_cache 0 --prewarm_primary_key_cache 1 2025-04-01 17:41:59 2025-04-01 17:41:59 Database: test_axigm1xv 2025-04-01 17:42:00 00558_aggregate_merge_totals_with_arenas: [ OK ] 0.58 sec. 2025-04-01 17:42:01 02525_different_engines_in_temporary_tables: [ OK ] 1.33 sec. 2025-04-01 17:42:01 01681_bloom_filter_nullable_column: [ OK ] 1.29 sec. 2025-04-01 17:42:03 00835_if_generic_case: [ OK ] 1.43 sec. 2025-04-01 17:42:03 02423_insert_stats_behaviour: [ OK ] 9.70 sec. 2025-04-01 17:42:03 02538_analyzer_create_table_as_select: [ OK ] 0.68 sec. 2025-04-01 17:42:04 03199_unbin_buffer_overflow: [ OK ] 13.56 sec. 2025-04-01 17:42:04 03003_prql_panic: [ OK ] 1.84 sec. 2025-04-01 17:42:04 01430_fix_any_rewrite_aliases: [ OK ] 0.53 sec. 2025-04-01 17:42:04 00143_number_classification_functions: [ OK ] 0.94 sec. 2025-04-01 17:42:05 03258_old_analyzer_const_expr_bug: [ OK ] 0.53 sec. 2025-04-01 17:42:05 03157_dynamic_type_json: [ OK ] 0.73 sec. 2025-04-01 17:42:05 00752_low_cardinality_lambda_argument: [ OK ] 1.03 sec. 2025-04-01 17:42:06 01825_new_type_json_bools: [ OK ] 0.70 sec. 2025-04-01 17:42:06 01891_partition_by_uuid: [ OK ] 0.64 sec. 2025-04-01 17:42:07 01414_freeze_does_not_prevent_alters: [ OK ] 1.12 sec. 2025-04-01 17:42:07 00506_union_distributed: [ OK ] 1.60 sec. 2025-04-01 17:42:08 01825_new_type_json_multiple_files: [ OK ] 8.45 sec. 2025-04-01 17:42:08 03166_optimize_row_order_during_insert: [ OK ] 1.23 sec. 2025-04-01 17:42:08 00299_stripe_log_multiple_inserts: [ OK ] 1.24 sec. 2025-04-01 17:42:08 00065_shard_float_literals_formatting: [ OK ] 0.68 sec. 2025-04-01 17:42:09 01620_fix_simple_state_arg_type: [ OK ] 0.89 sec. 2025-04-01 17:42:09 02981_variant_type_function: [ OK ] 1.33 sec. 2025-04-01 17:42:09 02811_invalid_embedded_rocksdb_create: [ OK ] 0.65 sec. 2025-04-01 17:42:10 02177_merge_optimize_aggregation_in_order: [ OK ] 0.84 sec. 2025-04-01 17:42:10 02012_compress_lz4: [ OK ] 2.33 sec. 2025-04-01 17:42:11 00809_add_days_segfault: [ OK ] 1.24 sec. 2025-04-01 17:42:11 02882_primary_key_index_in_function_different_types: [ OK ] 0.83 sec. 2025-04-01 17:42:12 00904_array_with_constant_2: [ OK ] 0.58 sec. 2025-04-01 17:42:12 03001_data_version_column: [ OK ] 0.94 sec. 2025-04-01 17:42:12 02477_exists_fuzz_43478: [ OK ] 0.58 sec. 2025-04-01 17:42:12 01493_table_function_null: [ OK ] 0.53 sec. 2025-04-01 17:42:13 00476_pretty_formats_and_widths: [ OK ] 0.63 sec. 2025-04-01 17:42:14 01085_simdjson_uint64: [ OK ] 0.59 sec. 2025-04-01 17:42:14 03032_storage_memory_modify_settings: [ OK ] 1.73 sec. 2025-04-01 17:42:15 03221_merge_profile_events: [ OK ] 14.16 sec. 2025-04-01 17:42:15 02718_cli_dashed_options_parsing: [ OK ] 4.95 sec. 2025-04-01 17:42:15 02752_space_function: [ OK ] 1.58 sec. 2025-04-01 17:42:15 03207_json_read_subcolumns_2_wide_merge_tree: [ SKIPPED ] 0.00 sec. 2025-04-01 17:42:15 Reason: not running for current build 2025-04-01 17:42:16 00974_full_outer_join: [ OK ] 0.70 sec. 2025-04-01 17:42:16 00098_f_union_all: [ OK ] 0.68 sec. 2025-04-01 17:42:17 03036_schema_inference_cache_s3_archives: [ OK ] 0.83 sec. 2025-04-01 17:42:17 03276_index_of_assume_sorted: [ OK ] 0.88 sec. 2025-04-01 17:42:17 01552_dict_fixedstring: [ OK ] 0.73 sec. 2025-04-01 17:42:18 01948_group_bitmap_and_or_xor_fix: [ OK ] 0.68 sec. 2025-04-01 17:42:18 00180_attach_materialized_view: [ OK ] 0.59 sec. 2025-04-01 17:42:18 02236_json_each_row_empty_map_schema_inference: [ OK ] 0.59 sec. 2025-04-01 17:42:19 03272_bad_aggregate_function: [ OK ] 0.73 sec. 2025-04-01 17:42:19 01825_new_type_json_9: [ OK ] 0.68 sec. 2025-04-01 17:42:19 02149_issue_32487: [ OK ] 0.53 sec. 2025-04-01 17:42:20 02680_illegal_type_of_filter_projection: [ OK ] 1.01 sec. 2025-04-01 17:42:21 02003_memory_limit_in_client: [ OK ] 69.11 sec. 2025-04-01 17:42:21 02491_part_log_has_table_uuid: [ OK ] 2.24 sec. 2025-04-01 17:42:22 02293_h3_hex_ring: [ OK ] 1.28 sec. 2025-04-01 17:42:22 03069_analyzer_with_alias_in_array_join: [ OK ] 0.60 sec. 2025-04-01 17:42:23 02474_extract_fixedstring_from_json: [ OK ] 1.28 sec. 2025-04-01 17:42:24 00914_replicate: [ OK ] 0.73 sec. 2025-04-01 17:42:35 02129_skip_quoted_fields: [ OK ] 20.19 sec. 2025-04-01 17:42:37 02562_native_tskv_default_for_omitted_fields: [ OK ] 12.54 sec. 2025-04-01 17:42:38 02346_fulltext_index_search: [ OK ] 40.30 sec. 2025-04-01 17:42:38 00831_quantile_weighted_parameter_check: [ OK ] 1.48 sec. 2025-04-01 17:42:38 00534_functions_bad_arguments8: [ SKIPPED ] 0.00 sec. 2025-04-01 17:42:38 Reason: not running for current build 2025-04-01 17:42:41 01077_mutations_index_consistency: [ OK ] 18.85 sec. 2025-04-01 17:42:41 02504_regexp_dictionary_ua_parser: [ OK ] 26.98 sec. 2025-04-01 17:42:43 01825_type_json_13: [ OK ] 7.05 sec. 2025-04-01 17:42:43 02842_largestTriangleThreeBuckets_aggregate_function: [ OK ] 4.38 sec. 2025-04-01 17:42:43 01034_unknown_qualified_column_in_join: [ OK ] 1.73 sec. 2025-04-01 17:42:46 00712_prewhere_with_final: [ OK ] 2.65 sec. 2025-04-01 17:42:46 02571_local_desc_abort_on_twitter_json: [ OK ] 2.94 sec. 2025-04-01 17:42:47 03009_format_show_database: [ OK ] 3.84 sec. 2025-04-01 17:42:47 00678_shard_funnel_window: [ OK ] 8.85 sec. 2025-04-01 17:42:49 00362_great_circle_distance: [ OK ] 1.40 sec. 2025-04-01 17:42:49 00568_empty_function_with_fixed_string: [ OK ] 1.58 sec. 2025-04-01 17:42:49 01273_arrow_stream: [ OK ] 41.63 sec. 2025-04-01 17:42:51 02841_not_ready_set_bug: [ OK ] 10.09 sec. 2025-04-01 17:42:53 02461_cancel_finish_race: [ OK ] 31.49 sec. 2025-04-01 17:42:53 02027_ngrams: [ OK ] 1.79 sec. 2025-04-01 17:42:53 02355_control_block_size_in_array_join: [ OK ] 3.83 sec. 2025-04-01 17:42:53 01890_state_of_state: [ OK ] 4.12 sec. 2025-04-01 17:42:53 02993_lazy_index_loading: [ OK ] 88.20 sec. 2025-04-01 17:42:54 02325_dates_schema_inference: [ OK ] 1.13 sec. 2025-04-01 17:42:54 01700_system_zookeeper_path_in: [ OK ] 0.83 sec. 2025-04-01 17:42:54 00460_vertical_and_totals_extremes: [ OK ] 0.59 sec. 2025-04-01 17:42:54 01784_parallel_formatting_memory: [ OK ] 1.08 sec. 2025-04-01 17:42:54 02932_query_settings_max_size_drop: [ OK ] 1.08 sec. 2025-04-01 17:42:54 02517_infer_uint64_in_case_of_int64_overflow: [ OK ] 8.70 sec. 2025-04-01 17:42:55 02567_native_type_conversions: [ OK ] 5.20 sec. 2025-04-01 17:42:55 03256_invalid_mutation_query: [ OK ] 1.08 sec. 2025-04-01 17:42:55 01845_add_testcase_for_arrayElement: [ OK ] 0.58 sec. 2025-04-01 17:42:56 01755_shard_pruning_with_literal: [ OK ] 0.63 sec. 2025-04-01 17:42:56 00712_prewhere_with_alias: [ OK ] 1.13 sec. 2025-04-01 17:42:56 01034_JSONCompactEachRow: [ OK ] 1.63 sec. 2025-04-01 17:42:56 01271_show_privileges: [ OK ] 0.53 sec. 2025-04-01 17:42:56 02751_protobuf_ipv6: [ OK ] 2.28 sec. 2025-04-01 17:42:59 02876_s3_cluster_schema_inference_names_with_spaces: [ OK ] 2.64 sec. 2025-04-01 17:42:59 02177_issue_31009: [ SKIPPED ] 0.00 sec. 2025-04-01 17:42:59 Reason: not running for current build 2025-04-01 17:42:59 02165_h3_edge_length_km: [ OK ] 3.00 sec. 2025-04-01 17:43:00 03281_dynamic_coalesce: [ OK ] 3.21 sec. 2025-04-01 17:43:01 01831_max_streams: [ OK ] 1.27 sec. 2025-04-01 17:43:02 01079_bit_operations_using_bitset: [ OK ] 2.45 sec. 2025-04-01 17:43:04 01940_totimezone_operator_monotonicity: [ OK ] 2.75 sec. 2025-04-01 17:43:06 02922_respect_nulls_Nullable: [ OK ] 5.81 sec. 2025-04-01 17:43:07 02479_analyzer_aggregation_crash: [ OK ] 2.00 sec. 2025-04-01 17:43:07 01660_join_or_all: [ OK ] 10.86 sec. 2025-04-01 17:43:08 03267_min_parts_to_merge_at_once: [ OK ] 13.43 sec. 2025-04-01 17:43:08 02883_zookeeper_finalize_stress: [ OK ] 12.56 sec. 2025-04-01 17:43:08 00853_join_with_nulls_crash: [ OK ] 2.08 sec. 2025-04-01 17:43:08 01683_dist_INSERT_block_structure_mismatch: [ OK ] 0.63 sec. 2025-04-01 17:43:09 01893_jit_aggregation_function_min_long: [ OK ] 7.05 sec. 2025-04-01 17:43:09 01079_alter_default_zookeeper_long: [ OK ] 2.19 sec. 2025-04-01 17:43:10 03279_pr_3_way_joins_left_first: [ OK ] 2.48 sec. 2025-04-01 17:43:12 01073_crlf_end_of_line: [ OK ] 1.25 sec. 2025-04-01 17:43:12 03208_array_of_json_read_subcolumns_2_compact_merge_tree: [ SKIPPED ] 0.00 sec. 2025-04-01 17:43:12 Reason: not running for current build 2025-04-01 17:43:13 02457_filesystem_function: [ OK ] 1.27 sec. 2025-04-01 17:43:15 01773_min_max_time_system_parts_datetime64: [ OK ] 1.45 sec. 2025-04-01 17:43:19 03038_recursive_cte_postgres_4: [ OK ] 3.42 sec. 2025-04-01 17:43:21 00938_dataset_test: [ OK ] 1.92 sec. 2025-04-01 17:43:24 03202_dynamic_null_map_subcolumn: [ OK ] 14.29 sec. 2025-04-01 17:43:25 00742_require_join_strictness: [ OK ] 3.74 sec. 2025-04-01 17:43:28 00975_json_hang: [ OK ] 19.26 sec. 2025-04-01 17:43:31 03199_queries_with_new_analyzer: [ OK ] 4.90 sec. 2025-04-01 17:43:32 02504_disallow_arrayjoin_in_mutations: [ OK ] 3.40 sec. 2025-04-01 17:43:32 03129_cte_with_final: [ OK ] 1.43 sec. 2025-04-01 17:43:33 00612_union_query_with_subquery: [ OK ] 1.07 sec. 2025-04-01 17:43:36 01463_resample_overflow: [ OK ] 2.96 sec. 2025-04-01 17:43:37 02125_lz4_compression_bug_JSONCompactEachRow: [ OK ] 42.98 sec. 2025-04-01 17:43:38 02163_operators: [ OK ] 1.27 sec. 2025-04-01 17:43:41 02255_broken_parts_chain_on_start: [ OK ] 15.26 sec. 2025-04-01 17:43:41 02366_kql_create_table: [ OK ] 2.71 sec. 2025-04-01 17:43:42 00503_cast_const_nullable: [ OK ] 0.59 sec. 2025-04-01 17:43:42 00991_system_parts_race_condition_long: [ OK ] 34.02 sec. 2025-04-01 17:43:42 02931_alter_materialized_view_query_inconsistent: [ OK ] 1.49 sec. 2025-04-01 17:43:43 01404_roundUpToPowerOfTwoOrZero_safety: [ OK ] 0.68 sec. 2025-04-01 17:43:43 01072_optimize_skip_unused_shards_const_expr_eval: [ OK ] 7.42 sec. 2025-04-01 17:43:43 01060_window_view_event_tumble_to_asc: [ OK ] 10.03 sec. 2025-04-01 17:43:44 00411_merge_tree_where_const_in_set: [ OK ] 0.78 sec. 2025-04-01 17:43:44 01753_mutate_table_predicated_with_table: [ OK ] 0.73 sec. 2025-04-01 17:43:45 02494_array_function_range: [ OK ] 0.68 sec. 2025-04-01 17:43:45 02176_toStartOfWeek_overflow_pruning: [ OK ] 0.83 sec. 2025-04-01 17:43:45 01621_clickhouse_compressor: [ OK ] 2.03 sec. 2025-04-01 17:43:45 01326_hostname_alias: [ OK ] 0.53 sec. 2025-04-01 17:43:46 01412_row_from_totals: [ OK ] 0.94 sec. 2025-04-01 17:43:47 01268_mv_scalars: [ OK ] 1.41 sec. 2025-04-01 17:43:48 01359_codeql: [ OK ] 0.48 sec. 2025-04-01 17:43:48 00814_parsing_ub: [ OK ] 0.53 sec. 2025-04-01 17:43:49 03246_json_simd_rapid_parsers: [ OK ] 3.44 sec. 2025-04-01 17:43:49 02815_range_dict_no_direct_join: [ OK ] 0.93 sec. 2025-04-01 17:43:50 03156_nullable_number_tips: [ OK ] 0.98 sec. 2025-04-01 17:43:50 02263_format_insert_settings: [ OK ] 8.20 sec. 2025-04-01 17:43:50 00358_from_string_complex_types: [ OK ] 0.53 sec. 2025-04-01 17:43:50 01460_line_as_string_format: [ OK ] 63.65 sec. 2025-04-01 17:43:51 02387_parse_date_as_datetime: [ OK ] 0.63 sec. 2025-04-01 17:43:51 01049_join_low_card_crash: [ OK ] 0.93 sec. 2025-04-01 17:43:51 01019_array_fill: [ OK ] 0.83 sec. 2025-04-01 17:43:51 00686_client_exit_code: [ OK ] 1.74 sec. 2025-04-01 17:43:52 00975_sample_prewhere_distributed: [ OK ] 0.73 sec. 2025-04-01 17:43:52 00296_url_parameters: [ OK ] 0.93 sec. 2025-04-01 17:43:53 01056_predicate_optimizer_bugs: [ OK ] 1.93 sec. 2025-04-01 17:43:54 02016_aggregation_spark_bar: [ OK ] 2.28 sec. 2025-04-01 17:43:55 02454_create_table_with_custom_disk: [ OK ] 1.14 sec. 2025-04-01 17:43:56 03234_check_grant: [ OK ] 4.59 sec. 2025-04-01 17:43:56 00623_truncate_table: [ OK ] 1.99 sec. 2025-04-01 17:43:57 03161_cnf_reduction: [ OK ] 1.73 sec. 2025-04-01 17:43:57 03150_infer_type_variant: [ OK ] 0.58 sec. 2025-04-01 17:43:58 00662_has_nullable: [ OK ] 0.83 sec. 2025-04-01 17:43:59 02122_parallel_formatting_TSV: [ OK ] 7.15 sec. 2025-04-01 17:44:00 01547_query_log_current_database: [ OK ] 4.29 sec. 2025-04-01 17:44:01 00417_kill_query: [ OK ] 5.30 sec. 2025-04-01 17:44:03 00926_adaptive_index_granularity_collapsing_merge_tree: [ OK ] 1.54 sec. 2025-04-01 17:44:04 00842_array_with_constant_overflow: [ OK ] 0.99 sec. 2025-04-01 17:44:06 02661_read_from_archive_tzst: [ OK ] 20.48 sec. 2025-04-01 17:44:06 01505_log_distributed_deadlock: [ OK ] 0.69 sec. 2025-04-01 17:44:07 01144_multiword_data_types: [ OK ] 0.79 sec. 2025-04-01 17:44:08 00351_select_distinct_arrays_tuples: [ OK ] 0.64 sec. 2025-04-01 17:44:09 03113_analyzer_not_found_column_in_block_2: [ OK ] 0.78 sec. 2025-04-01 17:44:10 03130_analyzer_self_join_group_by: [ OK ] 1.03 sec. 2025-04-01 17:44:10 01632_tinylog_read_write: [ OK ] 11.82 sec. 2025-04-01 17:44:11 02732_rename_after_processing: [ OK ] 6.65 sec. 2025-04-01 17:44:12 01702_rewrite_avg_for_algebraic_optimization: [ OK ] 1.03 sec. 2025-04-01 17:44:13 01272_suspicious_codecs: [ OK ] 2.99 sec. 2025-04-01 17:44:14 00338_replicate_array_of_strings: [ OK ] 1.04 sec. 2025-04-01 17:44:15 02271_replace_partition_many_tables: [ OK ] 31.51 sec. 2025-04-01 17:44:15 02235_add_part_offset_virtual_column: [ OK ] 15.67 sec. 2025-04-01 17:44:15 01440_big_int_shift: [ OK ] 0.68 sec. 2025-04-01 17:44:15 01640_distributed_async_insert_compression: [ OK ] 0.79 sec. 2025-04-01 17:44:16 00465_nullable_default: [ OK ] 0.65 sec. 2025-04-01 17:44:16 01837_cast_to_array_from_empty_array: [ OK ] 0.59 sec. 2025-04-01 17:44:16 02394_every_profile_event_must_have_documentation: [ OK ] 0.55 sec. 2025-04-01 17:44:17 00803_odbc_driver_2_format: [ OK ] 0.50 sec. 2025-04-01 17:44:17 00910_buffer_prewhere: [ OK ] 0.67 sec. 2025-04-01 17:44:18 01825_type_json_3: [ OK ] 2.79 sec. 2025-04-01 17:44:18 01276_system_licenses: [ OK ] 1.15 sec. 2025-04-01 17:44:18 00165_transform_non_const_default: [ OK ] 1.00 sec. 2025-04-01 17:44:19 01011_group_uniq_array_memsan: [ OK ] 0.61 sec. 2025-04-01 17:44:19 03013_position_const_start_pos: [ OK ] 0.53 sec. 2025-04-01 17:44:19 01720_engine_file_empty_if_not_exists: [ OK ] 1.55 sec. 2025-04-01 17:44:20 01038_dictionary_lifetime_min_zero_sec: [ OK ] 7.87 sec. 2025-04-01 17:44:20 01851_s2_to_geo: [ OK ] 0.59 sec. 2025-04-01 17:44:20 02902_select_subcolumns_from_engine_null: [ OK ] 0.65 sec. 2025-04-01 17:44:20 02206_clickhouse_local_use_database: [ OK ] 1.94 sec. 2025-04-01 17:44:21 02923_join_use_nulls_modulo: [ OK ] 0.95 sec. 2025-04-01 17:44:22 01790_dist_INSERT_block_structure_mismatch_types_and_names: [ OK ] 1.44 sec. 2025-04-01 17:44:23 02317_functions_with_nothing: [ OK ] 1.39 sec. 2025-04-01 17:44:24 01849_geoToS2: [ OK ] 2.36 sec. 2025-04-01 17:44:24 02475_or_function_alias_and_const_where: [ OK ] 0.63 sec. 2025-04-01 17:44:25 02967_fuzz_bad_cast: [ OK ] 1.19 sec. 2025-04-01 17:44:26 03084_analyzer_join_column_alias: [ OK ] 1.59 sec. 2025-04-01 17:44:26 01498_alter_column_storage_memory: [ OK ] 0.71 sec. 2025-04-01 17:44:26 00263_merge_aggregates_and_overflow: [ OK ] 1.15 sec. 2025-04-01 17:44:29 02813_seriesOutliersDetectTukey: [ OK ] 2.41 sec. 2025-04-01 17:44:29 01064_arrayROCAUC: [ OK ] 2.71 sec. 2025-04-01 17:44:31 00688_low_cardinality_syntax: [ OK ] 1.80 sec. 2025-04-01 17:44:31 01772_to_start_of_hour_align: [ OK ] 1.50 sec. 2025-04-01 17:44:32 02540_analyzer_matcher_alias_materialized_columns: [ OK ] 0.95 sec. 2025-04-01 17:44:32 00521_multidimensional: [ OK ] 1.39 sec. 2025-04-01 17:44:34 02990_format_select_from_explain: [ OK ] 1.43 sec. 2025-04-01 17:44:34 03273_group_by_in_order_still_used_when_group_by_key_doesnt_match_order_by_key: [ OK ] 2.04 sec. 2025-04-01 17:44:35 00751_low_cardinality_nullable_group_by: [ OK ] 34.38 sec. 2025-04-01 17:44:35 02245_s3_schema_desc: [ OK ] 1.18 sec. 2025-04-01 17:44:35 02933_group_by_memory_usage: [ OK ] 25.74 sec. 2025-04-01 17:44:36 02915_fpc_overflow: [ OK ] 1.84 sec. 2025-04-01 17:44:36 00950_test_gorilla_codec: [ OK ] 1.18 sec. 2025-04-01 17:44:36 01925_json_as_string_data_in_square_brackets: [ OK ] 0.64 sec. 2025-04-01 17:44:36 03227_dynamic_subcolumns_enumerate_streams: [ OK ] 0.63 sec. 2025-04-01 17:44:37 02354_with_statement_non_exist_column: [ OK ] 0.60 sec. 2025-04-01 17:44:37 02862_sorted_distinct_sparse_fix: [ OK ] 0.81 sec. 2025-04-01 17:44:37 01088_array_slice_of_aggregate_functions: [ OK ] 0.65 sec. 2025-04-01 17:44:37 02877_optimize_read_in_order_from_view: [ OK ] 17.49 sec. 2025-04-01 17:44:37 02455_extract_fixed_string_from_nested_json: [ OK ] 0.59 sec. 2025-04-01 17:44:37 01651_group_uniq_array_enum: [ OK ] 0.58 sec. 2025-04-01 17:44:38 02265_per_table_ttl_mutation_on_change: [ OK ] 1.39 sec. 2025-04-01 17:44:38 02963_single_value_destructor: [ OK ] 1.13 sec. 2025-04-01 17:44:38 01009_insert_select_data_loss: [ OK ] 0.63 sec. 2025-04-01 17:44:38 02476_fix_cast_parser_bug: [ OK ] 0.43 sec. 2025-04-01 17:44:38 00950_bad_alloc_when_truncate_join_storage: [ OK ] 0.48 sec. 2025-04-01 17:44:39 01259_datetime64_ubsan: [ OK ] 0.84 sec. 2025-04-01 17:44:39 02266_auto_add_nullable: [ OK ] 0.63 sec. 2025-04-01 17:44:39 01451_wrong_error_long_query: [ OK ] 1.78 sec. 2025-04-01 17:44:39 00284_external_aggregation_2: [ OK ] 91.78 sec. 2025-04-01 17:44:39 00609_distributed_with_case_when_then: [ OK ] 0.73 sec. 2025-04-01 17:44:39 03276_functions_to_subcolumns_lc: [ OK ] 0.63 sec. 2025-04-01 17:44:40 01906_bigint_accurate_cast_ubsan: [ OK ] 1.43 sec. 2025-04-01 17:44:41 02674_trivial_count_analyzer: [ OK ] 1.43 sec. 2025-04-01 17:44:41 02874_toDaysSinceYearZero: [ OK ] 1.92 sec. 2025-04-01 17:44:42 02002_system_table_with_tuple: [ OK ] 2.00 sec. 2025-04-01 17:44:42 00628_in_lambda_on_merge_table_bug: [ OK ] 0.88 sec. 2025-04-01 17:44:43 01714_alter_drop_version: [ OK ] 0.98 sec. 2025-04-01 17:44:44 01461_query_start_time_microseconds: [ OK ] 6.75 sec. 2025-04-01 17:44:44 02950_parallel_replicas_used_count: [ OK ] 24.20 sec. 2025-04-01 17:44:44 02943_variant_element: [ OK ] 0.78 sec. 2025-04-01 17:44:45 03043_group_array_result_is_expected: [ OK ] 0.63 sec. 2025-04-01 17:44:45 01690_quantilesTiming_ubsan: [ OK ] 0.48 sec. 2025-04-01 17:44:45 01081_demangle: [ OK ] 0.53 sec. 2025-04-01 17:44:45 00287_column_const_with_nan: [ OK ] 0.52 sec. 2025-04-01 17:44:45 01392_column_resolve: [ OK ] 0.68 sec. 2025-04-01 17:44:48 01825_new_type_json_11: [ OK ] 6.80 sec. 2025-04-01 17:44:49 00590_limit_by_column_removal: [ OK ] 0.58 sec. 2025-04-01 17:44:50 02908_Npy_files_caching: [ OK ] 4.69 sec. 2025-04-01 17:44:51 02276_full_sort_join_unsupported: [ OK ] 1.63 sec. 2025-04-01 17:44:51 01014_count_of_merges_metrics: [ OK ] 0.74 sec. 2025-04-01 17:44:51 02481_default_value_used_in_row_level_filter: [ OK ] 0.73 sec. 2025-04-01 17:44:52 02481_low_cardinality_with_short_circuit_functins: [ OK ] 0.83 sec. 2025-04-01 17:44:52 02946_literal_alias_misclassification: [ OK ] 0.63 sec. 2025-04-01 17:44:53 01055_prewhere_bugs: [ OK ] 0.83 sec. 2025-04-01 17:44:54 01646_rewrite_sum_if: [ OK ] 1.23 sec. 2025-04-01 17:44:57 02461_prewhere_row_level_policy_lightweight_delete: [ OK ] 4.94 sec. 2025-04-01 17:44:57 00844_join_lightee2: [ OK ] 0.78 sec. 2025-04-01 17:44:59 02225_hints_for_indeices: [ OK ] 4.39 sec. 2025-04-01 17:44:59 00647_select_numbers_with_offset: [ OK ] 0.53 sec. 2025-04-01 17:45:00 02971_limit_by_distributed: [ OK ] 0.83 sec. 2025-04-01 17:45:02 01520_client_print_query_id: [ OK ] 1.93 sec. 2025-04-01 17:45:03 00964_bloom_index_string_functions: [ OK ] 17.37 sec. 2025-04-01 17:45:03 00757_enum_defaults_const: [ OK ] 0.64 sec. 2025-04-01 17:45:04 01013_totals_without_aggregation: [ OK ] 0.89 sec. 2025-04-01 17:45:04 00875_join_right_nulls_ors: [ OK ] 1.03 sec. 2025-04-01 17:45:05 01880_remote_ipv6: [ OK ] 1.13 sec. 2025-04-01 17:45:05 01910_view_dictionary_check_refresh: [ OK ] 25.19 sec. 2025-04-01 17:45:05 02941_variant_type_2: [ OK ] 118.67 sec. 2025-04-01 17:45:05 02962_analyzer_constant_set: [ OK ] 0.58 sec. 2025-04-01 17:45:06 02366_kql_distinct: [ OK ] 0.73 sec. 2025-04-01 17:45:06 03007_column_nullable_uninitialzed_value: [ OK ] 0.53 sec. 2025-04-01 17:45:06 02711_trim_aliases: [ OK ] 0.63 sec. 2025-04-01 17:45:07 00668_compare_arrays_silviucpp: [ OK ] 0.58 sec. 2025-04-01 17:45:07 01328_bad_peephole_optimization: [ OK ] 0.53 sec. 2025-04-01 17:45:07 02711_server_uuid_macro: [ OK ] 1.03 sec. 2025-04-01 17:45:08 02337_join_analyze_stuck: [ OK ] 2.18 sec. 2025-04-01 17:45:08 02504_parse_datetime_best_effort_calebeaires: [ OK ] 0.58 sec. 2025-04-01 17:45:08 00072_in_types: [ OK ] 0.53 sec. 2025-04-01 17:45:09 01093_cyclic_defaults_filimonov: [ OK ] 0.68 sec. 2025-04-01 17:45:09 02931_size_virtual_column_use_structure_from_insertion_table: [ OK ] 1.83 sec. 2025-04-01 17:45:09 02418_keeper_map_keys_limit: [ OK ] 1.09 sec. 2025-04-01 17:45:10 00152_totals_in_subquery: [ OK ] 0.54 sec. 2025-04-01 17:45:10 01037_zookeeper_check_table_empty_pk: [ OK ] 0.83 sec. 2025-04-01 17:45:11 02243_make_date32: [ OK ] 1.98 sec. 2025-04-01 17:45:11 02999_analyzer_preimage_null: [ OK ] 0.63 sec. 2025-04-01 17:45:11 00340_squashing_insert_select: [ OK ] 32.61 sec. 2025-04-01 17:45:11 03001_bad_error_message_higher_order_functions: [ OK ] 2.08 sec. 2025-04-01 17:45:12 03199_fix_auc_tie_handling: [ OK ] 0.68 sec. 2025-04-01 17:45:13 02895_peak_memory_usage_http_headers_regression: [ OK ] 2.03 sec. 2025-04-01 17:45:14 01788_update_nested_type_subcolumn_check: [ OK ] 2.09 sec. 2025-04-01 17:45:14 02301_harmful_reexec: [ OK ] 2.33 sec. 2025-04-01 17:45:14 03156_group_concat: [ OK ] 2.68 sec. 2025-04-01 17:45:14 02049_lowcardinality_shortcircuit_crash: [ OK ] 0.68 sec. 2025-04-01 17:45:15 00802_daylight_saving_time_shift_backwards_at_midnight: [ OK ] 0.63 sec. 2025-04-01 17:45:15 01120_join_constants: [ OK ] 0.68 sec. 2025-04-01 17:45:16 01773_datetime64_add_ubsan: [ OK ] 0.78 sec. 2025-04-01 17:45:16 01821_join_table_mutation: [ OK ] 0.98 sec. 2025-04-01 17:45:16 01747_alter_partition_key_enum_zookeeper_long: [ OK ] 1.53 sec. 2025-04-01 17:45:16 01102_distributed_local_in_bug: [ OK ] 0.78 sec. 2025-04-01 17:45:17 01937_nested_chinese: [ OK ] 0.68 sec. 2025-04-01 17:45:17 03145_unicode_quotes: [ OK ] 0.58 sec. 2025-04-01 17:45:17 01122_totals_rollup_having_block_header: [ OK ] 0.58 sec. 2025-04-01 17:45:17 03033_final_undefined_last_mark: [ OK ] 0.44 sec. 2025-04-01 17:45:18 02150_replace_regexp_all_empty_match: [ OK ] 0.48 sec. 2025-04-01 17:45:18 00732_quorum_insert_simple_test_2_parts_zookeeper_long: [ OK ] 0.98 sec. 2025-04-01 17:45:18 02346_fulltext_index_bug47393: [ OK ] 0.78 sec. 2025-04-01 17:45:19 03233_dynamic_in_functions: [ OK ] 5.84 sec. 2025-04-01 17:45:19 00169_join_constant_keys: [ OK ] 0.95 sec. 2025-04-01 17:45:19 00534_functions_bad_arguments2: [ SKIPPED ] 0.00 sec. 2025-04-01 17:45:19 Reason: not running for current build 2025-04-01 17:45:20 02493_analyzer_sum_if_to_count_if: [ OK ] 1.13 sec. 2025-04-01 17:45:20 01720_union_distinct_with_limit: [ OK ] 0.58 sec. 2025-04-01 17:45:21 01419_merge_tree_settings_sanity_check: [ OK ] 1.08 sec. 2025-04-01 17:45:21 00217_shard_global_subquery_columns_with_same_name: [ OK ] 0.63 sec. 2025-04-01 17:45:21 03234_proto_complex_nested_repeated_noexception: [ OK ] 2.22 sec. 2025-04-01 17:45:22 03006_buffer_overflow_join: [ OK ] 0.58 sec. 2025-04-01 17:45:22 02985_minmax_index_aggregate_function: [ OK ] 0.98 sec. 2025-04-01 17:45:23 02591_bson_long_tuple: [ OK ] 0.58 sec. 2025-04-01 17:45:23 02473_multistep_split_prewhere: [ OK ] 38.07 sec. 2025-04-01 17:45:24 02764_index_analysis_fix: [ OK ] 0.63 sec. 2025-04-01 17:45:24 02226_low_cardinality_text_bloom_filter_index: [ OK ] 1.68 sec. 2025-04-01 17:45:25 01889_check_row_policy_defined_using_user_function: [ OK ] 9.16 sec. 2025-04-01 17:45:25 02112_delayed_clickhouse_client_with_queries_file: [ OK ] 2.08 sec. 2025-04-01 17:45:25 00758_array_reverse: [ OK ] 0.99 sec. 2025-04-01 17:45:26 01548_create_table_compound_column_format: [ OK ] 1.69 sec. 2025-04-01 17:45:28 00972_geohashesInBox: [ OK ] 3.94 sec. 2025-04-01 17:45:28 02128_apply_lambda_parsing: [ OK ] 0.59 sec. 2025-04-01 17:45:28 02461_alter_update_respect_part_column_type_bug: [ OK ] 3.64 sec. 2025-04-01 17:45:29 03040_dynamic_type_alters_1_wide_merge_tree: [ OK ] 2.70 sec. 2025-04-01 17:45:29 02810_fix_remove_dedundant_distinct_view: [ OK ] 0.94 sec. 2025-04-01 17:45:30 01892_setting_limit_offset_distributed: [ OK ] 0.99 sec. 2025-04-01 17:45:31 02150_index_hypothesis_race_long: [ OK ] 10.80 sec. 2025-04-01 17:45:31 01442_merge_detach_attach_long: [ OK ] 33.91 sec. 2025-04-01 17:45:32 02835_fuzz_remove_redundant_sorting: [ OK ] 2.24 sec. 2025-04-01 17:45:32 02884_duplicate_index_name: [ OK ] 0.58 sec. 2025-04-01 17:45:33 02316_cast_to_ip_address_default_column: [ OK ] 0.98 sec. 2025-04-01 17:45:33 01451_replicated_detach_drop_part_long: [ OK ] 2.04 sec. 2025-04-01 17:45:34 01280_unicode_whitespaces_lexer: [ OK ] 0.59 sec. 2025-04-01 17:45:35 02041_test_fuzzy_alter: [ OK ] 0.74 sec. 2025-04-01 17:45:35 01710_projection_optimize_group_by_function_keys: [ OK ] 0.68 sec. 2025-04-01 17:45:37 00100_subquery_table_identifier: [ OK ] 4.19 sec. 2025-04-01 17:45:42 01099_operators_date_and_timestamp: [ OK ] 6.09 sec. 2025-04-01 17:45:42 00953_constraints_operations: [ OK ] 10.50 sec. 2025-04-01 17:45:43 01580_column_const_comparision: [ OK ] 0.63 sec. 2025-04-01 17:45:43 01415_sticking_mutations: [ OK ] 39.42 sec. 2025-04-01 17:45:44 01021_only_tuple_columns: [ OK ] 2.24 sec. 2025-04-01 17:45:44 00726_length_aliases: [ OK ] 0.55 sec. 2025-04-01 17:45:45 03230_date_trunc_and_to_start_of_interval_on_date32: [ OK ] 1.50 sec. 2025-04-01 17:45:45 01199_url_functions_path_without_schema_yiurule: [ OK ] 0.58 sec. 2025-04-01 17:45:46 01880_materialized_view_to_table_type_check: [ OK ] 1.08 sec. 2025-04-01 17:45:46 03200_subcolumns_join_use_nulls: [ OK ] 0.79 sec. 2025-04-01 17:45:47 03152_analyzer_columns_list: [ OK ] 0.98 sec. 2025-04-01 17:45:47 01550_mutation_subquery: [ OK ] 0.68 sec. 2025-04-01 17:45:47 02477_age: [ OK ] 1.53 sec. 2025-04-01 17:45:48 03031_clickhouse_local_input: [ OK ] 4.64 sec. 2025-04-01 17:45:48 01085_extract_all_empty: [ OK ] 0.53 sec. 2025-04-01 17:45:49 00357_to_string_complex_types: [ OK ] 0.78 sec. 2025-04-01 17:45:49 03215_validate_type_in_alter_add_modify_column: [ OK ] 1.58 sec. 2025-04-01 17:45:50 00952_part_frozen_info: [ OK ] 0.89 sec. 2025-04-01 17:45:50 03062_analyzer_join_engine_missing_column: [ OK ] 0.68 sec. 2025-04-01 17:45:50 01663_quantile_weighted_overflow: [ OK ] 0.58 sec. 2025-04-01 17:45:51 03266_with_fill_staleness: [ OK ] 0.88 sec. 2025-04-01 17:45:51 01398_in_tuple_func: [ OK ] 0.69 sec. 2025-04-01 17:45:51 01710_projections_optimize_aggregation_in_order: [ OK ] 21.13 sec. 2025-04-01 17:45:52 00055_join_two_numbers: [ OK ] 0.63 sec. 2025-04-01 17:45:52 02267_insert_empty_data: [ OK ] 0.53 sec. 2025-04-01 17:45:53 03140_client_subsequent_external_tables: [ OK ] 1.93 sec. 2025-04-01 17:46:03 02841_join_filter_set_sparse: [ FAIL ] 1.08 sec. 2025-04-01 17:46:03 Reason: return code: 32 2025-04-01 17:46:03 Error on processing query: Code: 32. DB::Exception: Attempt to read after eof: while receiving packet from localhost:9000. (ATTEMPT_TO_READ_AFTER_EOF) (version 24.12.2.20221.altinityantalya (altinity build)) 2025-04-01 17:46:03 (query: SELECT countIf(ignore(*) == 0) FROM t1 JOIN t2 ON t1.s = t2.s;) 2025-04-01 17:46:03 , result: 2025-04-01 17:46:03 2025-04-01 17:46:03 2025-04-01 17:46:03 2025-04-01 17:46:03 stdout: 2025-04-01 17:46:03 2025-04-01 17:46:03 2025-04-01 17:46:03 Settings used in the test: --max_insert_threads 3 --group_by_two_level_threshold 482426 --group_by_two_level_threshold_bytes 50000000 --distributed_aggregation_memory_efficient 1 --fsync_metadata 0 --output_format_parallel_formatting 0 --input_format_parallel_parsing 1 --min_chunk_bytes_for_parallel_parsing 21376022 --max_read_buffer_size 635170 --prefer_localhost_replica 0 --max_block_size 75845 --max_joined_block_size_rows 54219 --max_threads 2 --optimize_append_index 0 --optimize_if_chain_to_multiif 1 --optimize_if_transform_strings_to_enum 1 --optimize_read_in_order 0 --optimize_or_like_chain 0 --optimize_substitute_columns 0 --enable_multiple_prewhere_read_steps 1 --read_in_order_two_level_merge_threshold 91 --optimize_aggregation_in_order 1 --aggregation_in_order_max_block_bytes 44115711 --use_uncompressed_cache 0 --min_bytes_to_use_direct_io 1 --min_bytes_to_use_mmap_io 1 --local_filesystem_read_method read --remote_filesystem_read_method threadpool --local_filesystem_read_prefetch 1 --filesystem_cache_segments_batch_size 10 --read_from_filesystem_cache_if_exists_otherwise_bypass_cache 0 --throw_on_error_from_cache_on_write_operations 0 --remote_filesystem_read_prefetch 0 --allow_prefetched_read_pool_for_remote_filesystem 0 --filesystem_prefetch_max_memory_usage 64Mi --filesystem_prefetches_limit 0 --filesystem_prefetch_min_bytes_for_single_read_task 1Mi --filesystem_prefetch_step_marks 0 --filesystem_prefetch_step_bytes 0 --compile_aggregate_expressions 0 --compile_sort_description 0 --merge_tree_coarse_index_granularity 22 --optimize_distinct_in_order 1 --max_bytes_before_external_sort 0 --max_bytes_before_external_group_by 0 --max_bytes_before_remerge_sort 988381737 --min_compress_block_size 1817107 --max_compress_block_size 284446 --merge_tree_compact_parts_min_granules_to_multibuffer_read 49 --optimize_sorting_by_input_stream_properties 0 --http_response_buffer_size 7804188 --http_wait_end_of_query False --enable_memory_bound_merging_of_aggregation_results 0 --min_count_to_compile_expression 0 --min_count_to_compile_aggregate_expression 0 --min_count_to_compile_sort_description 3 --session_timezone Africa/Juba --use_page_cache_for_disks_without_file_cache True --page_cache_inject_eviction False --merge_tree_read_split_ranges_into_intersecting_and_non_intersecting_injection_probability 0.3 --prefer_external_sort_block_bytes 1 --cross_join_min_rows_to_compress 100000000 --cross_join_min_bytes_to_compress 100000000 --min_external_table_block_size_bytes 1 --max_parsing_threads 1 --optimize_functions_to_subcolumns 0 --parallel_replicas_local_plan 0 --query_plan_join_swap_table true --output_format_native_write_json_as_string 1 --enable_vertical_final 0 2025-04-01 17:46:03 2025-04-01 17:46:03 MergeTree settings used in test: --ratio_of_defaults_for_sparse_serialization 1.0 --prefer_fetch_merged_part_size_threshold 6240349003 --vertical_merge_algorithm_min_rows_to_activate 1000000 --vertical_merge_algorithm_min_columns_to_activate 100 --allow_vertical_merges_from_compact_to_wide_parts 1 --min_merge_bytes_to_use_direct_io 3664758498 --index_granularity_bytes 330795 --merge_max_block_size 19699 --index_granularity 11246 --min_bytes_for_wide_part 1073741824 --marks_compress_block_size 35030 --primary_key_compress_block_size 34332 --replace_long_file_name_to_hash 0 --max_file_name_length 55 --min_bytes_for_full_part_storage 0 --compact_parts_max_bytes_to_buffer 510861828 --compact_parts_max_granules_to_buffer 175 --compact_parts_merge_max_bytes_to_prefetch_part 12289243 --cache_populated_by_fetch 1 --concurrent_part_removal_threshold 35 --old_parts_lifetime 480 --prewarm_mark_cache 0 --use_const_adaptive_granularity 1 --enable_index_granularity_compression 1 --use_primary_key_cache 1 --prewarm_primary_key_cache 0 2025-04-01 17:46:03 2025-04-01 17:46:03 Database: test_rnvzzqsk 2025-04-01 17:46:03 01671_aggregate_function_group_bitmap_data: [ FAIL ] 0.68 sec. 2025-04-01 17:46:03 Reason: return code: 210 2025-04-01 17:46:03 Code: 210. DB::NetException: Connection reset by peer, while reading from socket (peer: [::1]:9000, local: [::1]:58584): (localhost:9000, ::1, local address: [::1]:58584). (NETWORK_ERROR) 2025-04-01 17:46:03 2025-04-01 17:46:03 , result: 2025-04-01 17:46:03 2025-04-01 17:46:03 2025-04-01 17:46:03 2025-04-01 17:46:03 stdout: 2025-04-01 17:46:03 2025-04-01 17:46:03 2025-04-01 17:46:03 Settings used in the test: --max_insert_threads 2 --group_by_two_level_threshold 1 --group_by_two_level_threshold_bytes 50000000 --distributed_aggregation_memory_efficient 1 --fsync_metadata 0 --output_format_parallel_formatting 1 --input_format_parallel_parsing 1 --min_chunk_bytes_for_parallel_parsing 12201714 --max_read_buffer_size 865507 --prefer_localhost_replica 0 --max_block_size 97703 --max_joined_block_size_rows 29130 --max_threads 3 --optimize_append_index 1 --optimize_if_chain_to_multiif 0 --optimize_if_transform_strings_to_enum 0 --optimize_read_in_order 0 --optimize_or_like_chain 0 --optimize_substitute_columns 1 --enable_multiple_prewhere_read_steps 1 --read_in_order_two_level_merge_threshold 32 --optimize_aggregation_in_order 0 --aggregation_in_order_max_block_bytes 34195147 --use_uncompressed_cache 0 --min_bytes_to_use_direct_io 6725307069 --min_bytes_to_use_mmap_io 10737418240 --local_filesystem_read_method io_uring --remote_filesystem_read_method read --local_filesystem_read_prefetch 0 --filesystem_cache_segments_batch_size 1 --read_from_filesystem_cache_if_exists_otherwise_bypass_cache 1 --throw_on_error_from_cache_on_write_operations 1 --remote_filesystem_read_prefetch 0 --allow_prefetched_read_pool_for_remote_filesystem 0 --filesystem_prefetch_max_memory_usage 64Mi --filesystem_prefetches_limit 10 --filesystem_prefetch_min_bytes_for_single_read_task 8Mi --filesystem_prefetch_step_marks 0 --filesystem_prefetch_step_bytes 0 --compile_aggregate_expressions 1 --compile_sort_description 1 --merge_tree_coarse_index_granularity 22 --optimize_distinct_in_order 0 --max_bytes_before_external_sort 10737418240 --max_bytes_before_external_group_by 0 --max_bytes_before_remerge_sort 337055548 --min_compress_block_size 2434648 --max_compress_block_size 153032 --merge_tree_compact_parts_min_granules_to_multibuffer_read 8 --optimize_sorting_by_input_stream_properties 0 --http_response_buffer_size 1300899 --http_wait_end_of_query True --enable_memory_bound_merging_of_aggregation_results 1 --min_count_to_compile_expression 0 --min_count_to_compile_aggregate_expression 3 --min_count_to_compile_sort_description 0 --session_timezone Africa/Juba --use_page_cache_for_disks_without_file_cache False --page_cache_inject_eviction False --merge_tree_read_split_ranges_into_intersecting_and_non_intersecting_injection_probability 0.57 --prefer_external_sort_block_bytes 1 --cross_join_min_rows_to_compress 100000000 --cross_join_min_bytes_to_compress 100000000 --min_external_table_block_size_bytes 1 --max_parsing_threads 10 --optimize_functions_to_subcolumns 1 --parallel_replicas_local_plan 0 --query_plan_join_swap_table auto --output_format_native_write_json_as_string 0 --enable_vertical_final 0 2025-04-01 17:46:03 2025-04-01 17:46:03 MergeTree settings used in test: --ratio_of_defaults_for_sparse_serialization 1.0 --prefer_fetch_merged_part_size_threshold 1 --vertical_merge_algorithm_min_rows_to_activate 1 --vertical_merge_algorithm_min_columns_to_activate 1 --allow_vertical_merges_from_compact_to_wide_parts 1 --min_merge_bytes_to_use_direct_io 10737418240 --index_granularity_bytes 5894704 --merge_max_block_size 3456 --index_granularity 6361 --min_bytes_for_wide_part 0 --marks_compress_block_size 38760 --primary_key_compress_block_size 61354 --replace_long_file_name_to_hash 1 --max_file_name_length 46 --min_bytes_for_full_part_storage 0 --compact_parts_max_bytes_to_buffer 28684583 --compact_parts_max_granules_to_buffer 253 --compact_parts_merge_max_bytes_to_prefetch_part 4049434 --cache_populated_by_fetch 1 --concurrent_part_removal_threshold 100 --old_parts_lifetime 480 --prewarm_mark_cache 1 --use_const_adaptive_granularity 0 --enable_index_granularity_compression 0 --use_primary_key_cache 0 --prewarm_primary_key_cache 0 2025-04-01 17:46:03 2025-04-01 17:46:03 Database: test_umfejymz 2025-04-01 17:46:03 01681_hyperscan_debug_assertion: [ FAIL ] 11.92 sec. 2025-04-01 17:46:03 Reason: server died 2025-04-01 17:46:03 ConnectionRefusedError 2025-04-01 17:46:03 [Errno 111] Connection refused 2025-04-01 17:46:03 File "/usr/bin/clickhouse-test", line 1854, in run 2025-04-01 17:46:03 self._cleanup(result.status == TestStatus.OK) 2025-04-01 17:46:03 2025-04-01 17:46:03 File "/usr/bin/clickhouse-test", line 1917, in _cleanup 2025-04-01 17:46:03 self._cleanup_database(args, timeout) 2025-04-01 17:46:03 2025-04-01 17:46:03 File "/usr/bin/clickhouse-test", line 1951, in _cleanup_database 2025-04-01 17:46:03 clickhouse_execute( 2025-04-01 17:46:03 2025-04-01 17:46:03 File "/usr/bin/clickhouse-test", line 334, in clickhouse_execute 2025-04-01 17:46:03 return clickhouse_execute_http( 2025-04-01 17:46:03 2025-04-01 17:46:03 File "/usr/bin/clickhouse-test", line 315, in clickhouse_execute_http 2025-04-01 17:46:03 raise ex 2025-04-01 17:46:03 2025-04-01 17:46:03 File "/usr/bin/clickhouse-test", line 304, in clickhouse_execute_http 2025-04-01 17:46:03 client.request( 2025-04-01 17:46:03 2025-04-01 17:46:03 File "/usr/lib/python3.10/http/client.py", line 1283, in request 2025-04-01 17:46:03 self._send_request(method, url, body, headers, encode_chunked) 2025-04-01 17:46:03 2025-04-01 17:46:03 File "/usr/lib/python3.10/http/client.py", line 1329, in _send_request 2025-04-01 17:46:03 self.endheaders(body, encode_chunked=encode_chunked) 2025-04-01 17:46:03 2025-04-01 17:46:03 File "/usr/lib/python3.10/http/client.py", line 1278, in endheaders 2025-04-01 17:46:03 self._send_output(message_body, encode_chunked=encode_chunked) 2025-04-01 17:46:03 2025-04-01 17:46:03 File "/usr/lib/python3.10/http/client.py", line 1038, in _send_output 2025-04-01 17:46:03 self.send(msg) 2025-04-01 17:46:03 2025-04-01 17:46:03 Settings used in the test: --max_insert_threads 2 --group_by_two_level_threshold 1000000 --group_by_two_level_threshold_bytes 8003255 --distributed_aggregation_memory_efficient 1 --fsync_metadata 0 --output_format_parallel_formatting 1 --input_format_parallel_parsing 1 --min_chunk_bytes_for_parallel_parsing 12313890 --max_read_buffer_size 719803 --prefer_localhost_replica 0 --max_block_size 52379 --max_joined_block_size_rows 66586 --max_threads 3 --optimize_append_index 0 --optimize_if_chain_to_multiif 1 --optimize_if_transform_strings_to_enum 0 --optimize_read_in_order 0 --optimize_or_like_chain 1 --optimize_substitute_columns 0 --enable_multiple_prewhere_read_steps 1 --read_in_order_two_level_merge_threshold 39 --optimize_aggregation_in_order 0 --aggregation_in_order_max_block_bytes 17533293 --use_uncompressed_cache 1 --min_bytes_to_use_direct_io 10737418240 --min_bytes_to_use_mmap_io 10737418240 --local_filesystem_read_method mmap --remote_filesystem_read_method threadpool --local_filesystem_read_prefetch 0 --filesystem_cache_segments_batch_size 10 --read_from_filesystem_cache_if_exists_otherwise_bypass_cache 0 --throw_on_error_from_cache_on_write_operations 1 --remote_filesystem_read_prefetch 1 --allow_prefetched_read_pool_for_remote_filesystem 1 --filesystem_prefetch_max_memory_usage 128Mi --filesystem_prefetches_limit 10 --filesystem_prefetch_min_bytes_for_single_read_task 1Mi --filesystem_prefetch_step_marks 50 --filesystem_prefetch_step_bytes 100Mi --compile_aggregate_expressions 0 --compile_sort_description 1 --merge_tree_coarse_index_granularity 20 --optimize_distinct_in_order 1 --max_bytes_before_external_sort 0 --max_bytes_before_external_group_by 0 --max_bytes_before_remerge_sort 694971537 --min_compress_block_size 2979773 --max_compress_block_size 2457989 --merge_tree_compact_parts_min_granules_to_multibuffer_read 55 --optimize_sorting_by_input_stream_properties 0 --http_response_buffer_size 9866312 --http_wait_end_of_query False --enable_memory_bound_merging_of_aggregation_results 0 --min_count_to_compile_expression 0 --min_count_to_compile_aggregate_expression 0 --min_count_to_compile_sort_description 3 --session_timezone Mexico/BajaSur --use_page_cache_for_disks_without_file_cache True --page_cache_inject_eviction False --merge_tree_read_split_ranges_into_intersecting_and_non_intersecting_injection_probability 0.83 --prefer_external_sort_block_bytes 1 --cross_join_min_rows_to_compress 0 --cross_join_min_bytes_to_compress 100000000 --min_external_table_block_size_bytes 0 --max_parsing_threads 0 --optimize_functions_to_subcolumns 0 --parallel_replicas_local_plan 1 --query_plan_join_swap_table true --output_format_native_write_json_as_string 0 --enable_vertical_final 0 2025-04-01 17:46:03 2025-04-01 17:46:03 MergeTree settings used in test: --ratio_of_defaults_for_sparse_serialization 1.0 --prefer_fetch_merged_part_size_threshold 3998079087 --vertical_merge_algorithm_min_rows_to_activate 1000000 --vertical_merge_algorithm_min_columns_to_activate 100 --allow_vertical_merges_from_compact_to_wide_parts 0 --min_merge_bytes_to_use_direct_io 1 --index_granularity_bytes 31337897 --merge_max_block_size 3528 --index_granularity 42523 --min_bytes_for_wide_part 273598252 --marks_compress_block_size 50847 --primary_key_compress_block_size 47749 --replace_long_file_name_to_hash 1 --max_file_name_length 128 --min_bytes_for_full_part_storage 55418150 --compact_parts_max_bytes_to_buffer 450913845 --compact_parts_max_granules_to_buffer 168 --compact_parts_merge_max_bytes_to_prefetch_part 7133480 --cache_populated_by_fetch 1 --concurrent_part_removal_threshold 32 --old_parts_lifetime 10 --prewarm_mark_cache 0 --use_const_adaptive_granularity 0 --enable_index_granularity_compression 1 --use_primary_key_cache 1 --prewarm_primary_key_cache 0 2025-04-01 17:46:03 2025-04-01 17:46:03 Database: test_ofjicghs 2025-04-01 17:46:03 Process Process-9: 2025-04-01 17:46:03 Traceback (most recent call last): 2025-04-01 17:46:03 File "/usr/lib/python3.10/multiprocessing/process.py", line 314, in _bootstrap 2025-04-01 17:46:03 self.run() 2025-04-01 17:46:03 File "/usr/lib/python3.10/multiprocessing/process.py", line 108, in run 2025-04-01 17:46:03 self._target(*self._args, **self._kwargs) 2025-04-01 17:46:03 File "/usr/bin/clickhouse-test", line 2626, in run_tests_process 2025-04-01 17:46:03 return run_tests_array(*args, **kwargs) 2025-04-01 17:46:03 File "/usr/bin/clickhouse-test", line 2407, in run_tests_array 2025-04-01 17:46:03 stop_tests() 2025-04-01 17:46:03 File "/usr/bin/clickhouse-test", line 452, in stop_tests 2025-04-01 17:46:03 cleanup_child_processes(os.getpid()) 2025-04-01 17:46:03 File "/usr/bin/clickhouse-test", line 440, in cleanup_child_processes 2025-04-01 17:46:03 child_pgid = os.getpgid(child) 2025-04-01 17:46:03 ProcessLookupError: [Errno 3] No such process 2025-04-01 17:46:03 Child processes of 1396: 2025-04-01 17:46:03 41785 /bin/sh -c pgrep --parent 1396 -a 2025-04-01 17:46:04 02844_max_backup_bandwidth_s3: [ FAIL ] 6.00 sec. 2025-04-01 17:46:04 Reason: return code: 210 2025-04-01 17:46:04 Error on processing query: Code: 32. DB::Exception: Attempt to read after eof: while receiving packet from localhost:9000. (ATTEMPT_TO_READ_AFTER_EOF) (version 24.12.2.20221.altinityantalya (altinity build)) 2025-04-01 17:46:04 (query: SYSTEM FLUSH LOGS;) 2025-04-01 17:46:04 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) 2025-04-01 17:46:04 2025-04-01 17:46:04 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) 2025-04-01 17:46:04 2025-04-01 17:46:04 , result: 2025-04-01 17:46:04 2025-04-01 17:46:04 2025-04-01 17:46:04 2025-04-01 17:46:04 stdout: 2025-04-01 17:46:04 2025-04-01 17:46:04 2025-04-01 17:46:04 Settings used in the test: --max_insert_threads 2 --group_by_two_level_threshold 741683 --group_by_two_level_threshold_bytes 47159360 --distributed_aggregation_memory_efficient 1 --fsync_metadata 1 --output_format_parallel_formatting 0 --input_format_parallel_parsing 0 --min_chunk_bytes_for_parallel_parsing 12317057 --max_read_buffer_size 754288 --prefer_localhost_replica 1 --max_block_size 36740 --max_joined_block_size_rows 98736 --max_threads 1 --optimize_append_index 1 --optimize_if_chain_to_multiif 1 --optimize_if_transform_strings_to_enum 1 --optimize_read_in_order 1 --optimize_or_like_chain 1 --optimize_substitute_columns 0 --enable_multiple_prewhere_read_steps 1 --read_in_order_two_level_merge_threshold 25 --optimize_aggregation_in_order 1 --aggregation_in_order_max_block_bytes 36192420 --use_uncompressed_cache 1 --min_bytes_to_use_direct_io 6541293247 --min_bytes_to_use_mmap_io 10737418240 --local_filesystem_read_method pread_threadpool --remote_filesystem_read_method read --local_filesystem_read_prefetch 1 --filesystem_cache_segments_batch_size 50 --read_from_filesystem_cache_if_exists_otherwise_bypass_cache 1 --throw_on_error_from_cache_on_write_operations 1 --remote_filesystem_read_prefetch 0 --allow_prefetched_read_pool_for_remote_filesystem 0 --filesystem_prefetch_max_memory_usage 128Mi --filesystem_prefetches_limit 10 --filesystem_prefetch_min_bytes_for_single_read_task 16Mi --filesystem_prefetch_step_marks 0 --filesystem_prefetch_step_bytes 0 --compile_aggregate_expressions 1 --compile_sort_description 1 --merge_tree_coarse_index_granularity 21 --optimize_distinct_in_order 1 --max_bytes_before_external_sort 5244846134 --max_bytes_before_external_group_by 10737418240 --max_bytes_before_remerge_sort 1189983565 --min_compress_block_size 3063502 --max_compress_block_size 1838624 --merge_tree_compact_parts_min_granules_to_multibuffer_read 42 --optimize_sorting_by_input_stream_properties 1 --http_response_buffer_size 8139777 --http_wait_end_of_query True --enable_memory_bound_merging_of_aggregation_results 0 --min_count_to_compile_expression 3 --min_count_to_compile_aggregate_expression 0 --min_count_to_compile_sort_description 3 --session_timezone Mexico/BajaSur --use_page_cache_for_disks_without_file_cache True --page_cache_inject_eviction False --merge_tree_read_split_ranges_into_intersecting_and_non_intersecting_injection_probability 0.65 --prefer_external_sort_block_bytes 1 --cross_join_min_rows_to_compress 0 --cross_join_min_bytes_to_compress 0 --min_external_table_block_size_bytes 1 --max_parsing_threads 1 --optimize_functions_to_subcolumns 0 --parallel_replicas_local_plan 1 --query_plan_join_swap_table auto --output_format_native_write_json_as_string 1 --enable_vertical_final 1 2025-04-01 17:46:04 2025-04-01 17:46:04 MergeTree settings used in test: --ratio_of_defaults_for_sparse_serialization 1.0 --prefer_fetch_merged_part_size_threshold 3438478178 --vertical_merge_algorithm_min_rows_to_activate 1000000 --vertical_merge_algorithm_min_columns_to_activate 1 --allow_vertical_merges_from_compact_to_wide_parts 0 --min_merge_bytes_to_use_direct_io 10737418240 --index_granularity_bytes 17218000 --merge_max_block_size 4659 --index_granularity 39897 --min_bytes_for_wide_part 1073741824 --marks_compress_block_size 69552 --primary_key_compress_block_size 91799 --replace_long_file_name_to_hash 0 --max_file_name_length 128 --min_bytes_for_full_part_storage 342185656 --compact_parts_max_bytes_to_buffer 263592454 --compact_parts_max_granules_to_buffer 1 --compact_parts_merge_max_bytes_to_prefetch_part 6838531 --cache_populated_by_fetch 1 --concurrent_part_removal_threshold 1 --old_parts_lifetime 222 --prewarm_mark_cache 0 --use_const_adaptive_granularity 1 --enable_index_granularity_compression 1 --use_primary_key_cache 1 --prewarm_primary_key_cache 1 2025-04-01 17:46:04 2025-04-01 17:46:04 Database: test_89ntz7bn 2025-04-01 17:46:04 01903_correct_block_size_prediction_with_default: [ FAIL ] 73.22 sec. 2025-04-01 17:46:04 Reason: return code: 210 2025-04-01 17:46:04 Error on processing query: Code: 32. DB::Exception: Attempt to read after eof: while receiving packet from localhost:9000. (ATTEMPT_TO_READ_AFTER_EOF) (version 24.12.2.20221.altinityantalya (altinity build)) 2025-04-01 17:46:04 (query: SELECT uniq(15Id) FROM test_extract PREWHERE 15Id < 4 SETTINGS max_threads=1) 2025-04-01 17:46:04 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) 2025-04-01 17:46:04 2025-04-01 17:46:04 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) 2025-04-01 17:46:04 2025-04-01 17:46:04 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) 2025-04-01 17:46:04 2025-04-01 17:46:04 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) 2025-04-01 17:46:04 2025-04-01 17:46:04 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) 2025-04-01 17:46:04 2025-04-01 17:46:04 , result: 2025-04-01 17:46:04 2025-04-01 17:46:04 8 2025-04-01 17:46:04 8 2025-04-01 17:46:04 1 2025-04-01 17:46:04 2025-04-01 17:46:04 stdout: 2025-04-01 17:46:04 8 2025-04-01 17:46:04 8 2025-04-01 17:46:04 1 2025-04-01 17:46:04 2025-04-01 17:46:04 2025-04-01 17:46:04 2025-04-01 17:46:04 Database: test_ofr5rq7t 2025-04-01 17:46:05 03036_test_parquet_bloom_filter_push_down: [ FAIL ] 17.84 sec. 2025-04-01 17:46:05 Reason: having stderror: 2025-04-01 17:46:05 Code: 210. DB::NetException: Connection reset by peer, while reading from socket (peer: [::1]:9000, local: [::1]:58606): (localhost:9000, ::1, local address: [::1]:58606). (NETWORK_ERROR) 2025-04-01 17:46:05 2025-04-01 17:46:05 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) 2025-04-01 17:46:05 2025-04-01 17:46:05 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) 2025-04-01 17:46:05 2025-04-01 17:46:05 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) 2025-04-01 17:46:05 2025-04-01 17:46:05 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) 2025-04-01 17:46:05 2025-04-01 17:46:05 2025-04-01 17:46:05 stdout: 2025-04-01 17:46:05 1000 2025-04-01 17:46:05 bloom filter is off, all row groups should be read 2025-04-01 17:46:05 expect rows_read = select count() 2025-04-01 17:46:05 { 2025-04-01 17:46:05 "data": [ 2025-04-01 17:46:05 { 2025-04-01 17:46:05 "string": "AZSR", 2025-04-01 17:46:05 "flba": "WNMM" 2025-04-01 17:46:05 }, 2025-04-01 17:46:05 { 2025-04-01 17:46:05 "string": "PFJH", 2025-04-01 17:46:05 "flba": "GKJC" 2025-04-01 17:46:05 } 2025-04-01 17:46:05 ], 2025-04-01 17:46:05 "rows": 2, 2025-04-01 17:46:05 "statistics": { 2025-04-01 17:46:05 "rows_read": 1000, 2025-04-01 17:46:05 "bytes_read": 47419 2025-04-01 17:46:05 } 2025-04-01 17:46:05 } 2025-04-01 17:46:05 bloom filter is on, some row groups should be skipped 2025-04-01 17:46:05 expect rows_read much less than select count() 2025-04-01 17:46:05 { 2025-04-01 17:46:05 "data": [ 2025-04-01 17:46:05 { 2025-04-01 17:46:05 "string": "AZSR", 2025-04-01 17:46:05 "flba": "WNMM" 2025-04-01 17:46:05 }, 2025-04-01 17:46:05 { 2025-04-01 17:46:05 "string": "PFJH", 2025-04-01 17:46:05 "flba": "GKJC" 2025-04-01 17:46:05 } 2025-04-01 17:46:05 ], 2025-04-01 17:46:05 "rows": 2, 2025-04-01 17:46:05 "statistics": { 2025-04-01 17:46:05 "rows_read": 464, 2025-04-01 17:46:05 "bytes_read": 21703 2025-04-01 17:46:05 } 2025-04-01 17:46:05 } 2025-04-01 17:46:05 bloom filter is on, but where predicate contains data from 2 row groups out of 3. 2025-04-01 17:46:05 Rows read should be less than select count, but greater than previous selects 2025-04-01 17:46:05 { 2025-04-01 17:46:05 "data": [ 2025-04-01 17:46:05 { 2025-04-01 17:46:05 "string": "PFJH", 2025-04-01 17:46:05 "flba": "GKJC" 2025-04-01 17:46:05 }, 2025-04-01 17:46:05 { 2025-04-01 17:46:05 "string": "ZHZK", 2025-04-01 17:46:05 "flba": "HRWD" 2025-04-01 17:46:05 } 2025-04-01 17:46:05 ], 2025-04-01 17:46:05 "rows": 2, 2025-04-01 17:46:05 "statistics": { 2025-04-01 17:46:05 "rows_read": 536, 2025-04-01 17:46:05 "bytes_read": 25708 2025-04-01 17:46:05 } 2025-04-01 17:46:05 } 2025-04-01 17:46:05 bloom filter is on, but where predicate contains data from all row groups 2025-04-01 17:46:05 expect rows_read = select count() 2025-04-01 17:46:05 { 2025-04-01 17:46:05 "data": [ 2025-04-01 17:46:05 { 2025-04-01 17:46:05 "string": "PFJH", 2025-04-01 17:46:05 "flba": "GKJC" 2025-04-01 17:46:05 }, 2025-04-01 17:46:05 { 2025-04-01 17:46:05 "string": "OKAI", 2025-04-01 17:46:05 "flba": "UXGT" 2025-04-01 17:46:05 }, 2025-04-01 17:46:05 { 2025-04-01 17:46:05 "string": "ZHZK", 2025-04-01 17:46:05 "flba": "HRWD" 2025-04-01 17:46:05 } 2025-04-01 17:46:05 ], 2025-04-01 17:46:05 "rows": 3, 2025-04-01 17:46:05 "statistics": { 2025-04-01 17:46:05 "rows_read": 1000, 2025-04-01 17:46:05 "bytes_read": 47419 2025-04-01 17:46:05 } 2025-04-01 17:46:05 } 2025-04-01 17:46:05 IN check 2025-04-01 17:46:05 { 2025-04-01 17:46:05 "data": [ 2025-04-01 17:46:05 { 2025-04-01 17:46:05 "string": "PFJH", 2025-04-01 17:46:05 "flba": "GKJC" 2025-04-01 17:46:05 }, 2025-04-01 17:46:05 { 2025-04-01 17:46:05 "string": "ZHZK", 2025-04-01 17:46:05 "flba": "HRWD" 2025-04-01 17:46:05 } 2025-04-01 17:46:05 ], 2025-04-01 17:46:05 "rows": 2, 2025-04-01 17:46:05 "statistics": { 2025-04-01 17:46:05 "rows_read": 536, 2025-04-01 17:46:05 "bytes_read": 25708 2025-04-01 17:46:05 } 2025-04-01 17:46:05 } 2025-04-01 17:46:05 tuple in case, bf is off. 2025-04-01 17:46:05 { 2025-04-01 17:46:05 "data": [ 2025-04-01 17:46:05 { 2025-04-01 17:46:05 "string": "PFJH", 2025-04-01 17:46:05 "flba": "GKJC" 2025-04-01 17:46:05 } 2025-04-01 17:46:05 ], 2025-04-01 17:46:05 "rows": 1, 2025-04-01 17:46:05 "statistics": { 2025-04-01 17:46:05 "rows_read": 1000, 2025-04-01 17:46:05 "bytes_read": 47419 2025-04-01 17:46:05 } 2025-04-01 17:46:05 } 2025-04-01 17:46:05 tuple in case, bf is on. 2025-04-01 17:46:05 { 2025-04-01 17:46:05 "data": [ 2025-04-01 17:46:05 { 2025-04-01 17:46:05 "string": "PFJH", 2025-04-01 17:46:05 "flba": "GKJC" 2025-04-01 17:46:05 } 2025-04-01 17:46:05 ], 2025-04-01 17:46:05 "rows": 1, 2025-04-01 17:46:05 "statistics": { 2025-04-01 17:46:05 "rows_read": 464, 2025-04-01 17:46:05 "bytes_read": 21703 2025-04-01 17:46:05 } 2025-04-01 17:46:05 } 2025-04-01 17:46:05 complex tuple in case, bf is off 2025-04-01 17:46:05 { 2025-04-01 17:46:05 "data": [ 2025-04-01 17:46:05 { 2025-04-01 17:46:05 "string": "PFJH", 2025-04-01 17:46:05 "flba": "GKJC" 2025-04-01 17:46:05 } 2025-04-01 17:46:05 ], 2025-04-01 17:46:05 "rows": 1, 2025-04-01 17:46:05 "statistics": { 2025-04-01 17:46:05 "rows_read": 1000, 2025-04-01 17:46:05 "bytes_read": 47419 2025-04-01 17:46:05 } 2025-04-01 17:46:05 } 2025-04-01 17:46:05 complex tuple in case, bf is on 2025-04-01 17:46:05 { 2025-04-01 17:46:05 "data": [ 2025-04-01 17:46:05 { 2025-04-01 17:46:05 "string": "PFJH", 2025-04-01 17:46:05 "flba": "GKJC" 2025-04-01 17:46:05 } 2025-04-01 17:46:05 ], 2025-04-01 17:46:05 "rows": 1, 2025-04-01 17:46:05 "statistics": { 2025-04-01 17:46:05 "rows_read": 464, 2025-04-01 17:46:05 "bytes_read": 21703 2025-04-01 17:46:05 } 2025-04-01 17:46:05 } 2025-04-01 17:46:05 complex tuple in case, bf is on. Non existent 2025-04-01 17:46:05 { 2025-04-01 17:46:05 "data": [], 2025-04-01 17:46:05 "rows": 0, 2025-04-01 17:46:05 "statistics": { 2025-04-01 17:46:05 "rows_read": 0, 2025-04-01 17:46:05 "bytes_read": 0 2025-04-01 17:46:05 } 2025-04-01 17:46:05 } 2025-04-01 17:46:05 Bloom filter for json column. BF is off 2025-04-01 17:46:05 { 2025-04-01 17:46:05 "data": [ 2025-04-01 17:46:05 { 2025-04-01 17:46:05 "json": "{\"key\":38, \"value\":\"NXONM\"}" 2025-04-01 17:46:05 } 2025-04-01 17:46:05 ], 2025-04-01 17:46:05 "rows": 1, 2025-04-01 17:46:05 "statistics": { 2025-04-01 17:46:05 "rows_read": 1000, 2025-04-01 17:46:05 "bytes_read": 47419 2025-04-01 17:46:05 } 2025-04-01 17:46:05 } 2025-04-01 17:46:05 Bloom filter for json column. BF is on 2025-04-01 17:46:05 { 2025-04-01 17:46:05 "data": [ 2025-04-01 17:46:05 { 2025-04-01 17:46:05 "json": "{\"key\":38, \"value\":\"NXONM\"}" 2025-04-01 17:46:05 } 2025-04-01 17:46:05 ], 2025-04-01 17:46:05 "rows": 1, 2025-04-01 17:46:05 "statistics": { 2025-04-01 17:46:05 "rows_read": 72, 2025-04-01 17:46:05 "bytes_read": 4005 2025-04-01 17:46:05 } 2025-04-01 17:46:05 } 2025-04-01 17:46:05 Bloom filter for ipv4 column. BF is off 2025-04-01 17:46:05 { 2025-04-01 17:46:05 "data": [ 2025-04-01 17:46:05 { 2025-04-01 17:46:05 "json": "{\"key\":38, \"value\":\"NXONM\"}" 2025-04-01 17:46:05 } 2025-04-01 17:46:05 ], 2025-04-01 17:46:05 "rows": 1, 2025-04-01 17:46:05 "statistics": { 2025-04-01 17:46:05 "rows_read": 1000, 2025-04-01 17:46:05 "bytes_read": 47419 2025-04-01 17:46:05 } 2025-04-01 17:46:05 } 2025-04-01 17:46:05 Bloom filter for ipv4 column. BF is on 2025-04-01 17:46:05 { 2025-04-01 17:46:05 "data": [ 2025-04-01 17:46:05 { 2025-04-01 17:46:05 "json": "{\"key\":38, \"value\":\"NXONM\"}" 2025-04-01 17:46:05 } 2025-04-01 17:46:05 ], 2025-04-01 17:46:05 "rows": 1, 2025-04-01 17:46:05 "statistics": { 2025-04-01 17:46:05 "rows_read": 72, 2025-04-01 17:46:05 "bytes_read": 4005 2025-04-01 17:46:05 } 2025-04-01 17:46:05 } 2025-04-01 17:46:05 Bloom filter for ipv4 column. BF is on. Specified in the schema 2025-04-01 17:46:05 { 2025-04-01 17:46:05 "data": [ 2025-04-01 17:46:05 { 2025-04-01 17:46:05 "ipv4": "0.0.1.143" 2025-04-01 17:46:05 } 2025-04-01 17:46:05 ], 2025-04-01 17:46:05 "rows": 1, 2025-04-01 17:46:05 "statistics": { 2025-04-01 17:46:05 "rows_read": 72, 2025-04-01 17:46:05 "bytes_read": 4005 2025-04-01 17:46:05 } 2025-04-01 17:46:05 } 2025-04-01 17:46:05 Bloom filter on 64 bit column read as ipv4. We explicitly deny it, should read all rg 2025-04-01 17:46:05 { 2025-04-01 17:46:05 "data": [ 2025-04-01 17:46:05 { 2025-04-01 17:46:05 "uint64_logical": "22.230.220.164" 2025-04-01 17:46:05 } 2025-04-01 17:46:05 ], 2025-04-01 17:46:05 "rows": 1, 2025-04-01 17:46:05 "statistics": { 2025-04-01 17:46:05 "rows_read": 1000, 2025-04-01 17:46:05 "bytes_read": 47419 2025-04-01 17:46:05 } 2025-04-01 17:46:05 } 2025-04-01 17:46:05 BF off for parquet uint64 logical type. Should read everything 2025-04-01 17:46:05 { 2025-04-01 17:46:05 "data": [ 2025-04-01 17:46:05 { 2025-04-01 17:46:05 "json": "{\"key\":683, \"value\":\"YKCPD\"}" 2025-04-01 17:46:05 } 2025-04-01 17:46:05 ], 2025-04-01 17:46:05 "rows": 1, 2025-04-01 17:46:05 "statistics": { 2025-04-01 17:46:05 "rows_read": 1000, 2025-04-01 17:46:05 "bytes_read": 47419 2025-04-01 17:46:05 } 2025-04-01 17:46:05 } 2025-04-01 17:46:05 BF on for parquet uint64 logical type. Uint64 is stored as a signed int 64, but with logical annotation. Make sure a value greater than int64 can be queried 2025-04-01 17:46:05 { 2025-04-01 17:46:05 "data": [ 2025-04-01 17:46:05 { 2025-04-01 17:46:05 "json": "{\"key\":683, \"value\":\"YKCPD\"}" 2025-04-01 17:46:05 } 2025-04-01 17:46:05 ], 2025-04-01 17:46:05 "rows": 1, 2025-04-01 17:46:05 "statistics": { 2025-04-01 17:46:05 "rows_read": 464, 2025-04-01 17:46:05 "bytes_read": 21711 2025-04-01 17:46:05 } 2025-04-01 17:46:05 } 2025-04-01 17:46:05 Uint16 is stored as physical type int32 with bidwidth = 16 and sign = false. Make sure a value greater than int16 can be queried. BF is on. 2025-04-01 17:46:05 BF off for parquet int8 logical type. Should read everything 2025-04-01 17:46:05 BF on for parquet int8 logical type. Should skip row groups 2025-04-01 17:46:05 Invalid column conversion with in operation. String type can not be hashed against parquet int64 physical type. Should read everything 2025-04-01 17:46:05 Transformations on key column shall not be allowed. Should read everything 2025-04-01 17:46:05 2025-04-01 17:46:05 Settings used in the test: --max_insert_threads 1 --group_by_two_level_threshold 176251 --group_by_two_level_threshold_bytes 50000000 --distributed_aggregation_memory_efficient 0 --fsync_metadata 0 --output_format_parallel_formatting 0 --input_format_parallel_parsing 0 --min_chunk_bytes_for_parallel_parsing 6090155 --max_read_buffer_size 661397 --prefer_localhost_replica 0 --max_block_size 37458 --max_joined_block_size_rows 59705 --max_threads 1 --optimize_append_index 0 --optimize_if_chain_to_multiif 1 --optimize_if_transform_strings_to_enum 1 --optimize_read_in_order 0 --optimize_or_like_chain 1 --optimize_substitute_columns 1 --enable_multiple_prewhere_read_steps 1 --read_in_order_two_level_merge_threshold 36 --optimize_aggregation_in_order 1 --aggregation_in_order_max_block_bytes 777581 --use_uncompressed_cache 0 --min_bytes_to_use_direct_io 8378955596 --min_bytes_to_use_mmap_io 10737418240 --local_filesystem_read_method read --remote_filesystem_read_method threadpool --local_filesystem_read_prefetch 0 --filesystem_cache_segments_batch_size 2 --read_from_filesystem_cache_if_exists_otherwise_bypass_cache 0 --throw_on_error_from_cache_on_write_operations 0 --remote_filesystem_read_prefetch 1 --allow_prefetched_read_pool_for_remote_filesystem 1 --filesystem_prefetch_max_memory_usage 128Mi --filesystem_prefetches_limit 0 --filesystem_prefetch_min_bytes_for_single_read_task 8Mi --filesystem_prefetch_step_marks 0 --filesystem_prefetch_step_bytes 100Mi --compile_aggregate_expressions 1 --compile_sort_description 1 --merge_tree_coarse_index_granularity 19 --optimize_distinct_in_order 1 --max_bytes_before_external_sort 10737418240 --max_bytes_before_external_group_by 10737418240 --max_bytes_before_remerge_sort 433165382 --min_compress_block_size 1014462 --max_compress_block_size 1822062 --merge_tree_compact_parts_min_granules_to_multibuffer_read 39 --optimize_sorting_by_input_stream_properties 0 --http_response_buffer_size 2615188 --http_wait_end_of_query True --enable_memory_bound_merging_of_aggregation_results 0 --min_count_to_compile_expression 0 --min_count_to_compile_aggregate_expression 3 --min_count_to_compile_sort_description 3 --session_timezone Mexico/BajaSur --use_page_cache_for_disks_without_file_cache True --page_cache_inject_eviction False --merge_tree_read_split_ranges_into_intersecting_and_non_intersecting_injection_probability 0.67 --prefer_external_sort_block_bytes 1 --cross_join_min_rows_to_compress 1 --cross_join_min_bytes_to_compress 0 --min_external_table_block_size_bytes 0 --max_parsing_threads 1 --optimize_functions_to_subcolumns 0 --parallel_replicas_local_plan 1 --query_plan_join_swap_table false --output_format_native_write_json_as_string 1 --enable_vertical_final 0 2025-04-01 17:46:05 2025-04-01 17:46:05 MergeTree settings used in test: --ratio_of_defaults_for_sparse_serialization 1.0 --prefer_fetch_merged_part_size_threshold 10737418240 --vertical_merge_algorithm_min_rows_to_activate 1000000 --vertical_merge_algorithm_min_columns_to_activate 100 --allow_vertical_merges_from_compact_to_wide_parts 1 --min_merge_bytes_to_use_direct_io 10737418240 --index_granularity_bytes 670404 --merge_max_block_size 7911 --index_granularity 39340 --min_bytes_for_wide_part 0 --marks_compress_block_size 81766 --primary_key_compress_block_size 8481 --replace_long_file_name_to_hash 1 --max_file_name_length 119 --min_bytes_for_full_part_storage 536870912 --compact_parts_max_bytes_to_buffer 261665137 --compact_parts_max_granules_to_buffer 204 --compact_parts_merge_max_bytes_to_prefetch_part 14650012 --cache_populated_by_fetch 0 --concurrent_part_removal_threshold 41 --old_parts_lifetime 213 --prewarm_mark_cache 1 --use_const_adaptive_granularity 1 --enable_index_granularity_compression 0 --use_primary_key_cache 1 --prewarm_primary_key_cache 0 2025-04-01 17:46:05 2025-04-01 17:46:05 Database: test_jypz7w9c 2025-04-01 17:46:07 02317_distinct_in_order_optimization_explain: [ FAIL ] 29.27 sec. 2025-04-01 17:46:07 Reason: return code: 210 2025-04-01 17:46:07 Code: 210. DB::NetException: Connection reset by peer, while reading from socket (peer: [::1]:9000, local: [::1]:58594): (localhost:9000, ::1, local address: [::1]:58594). (NETWORK_ERROR) 2025-04-01 17:46:07 2025-04-01 17:46:07 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) 2025-04-01 17:46:07 2025-04-01 17:46:07 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) 2025-04-01 17:46:07 2025-04-01 17:46:07 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) 2025-04-01 17:46:07 2025-04-01 17:46:07 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) 2025-04-01 17:46:07 2025-04-01 17:46:07 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) 2025-04-01 17:46:07 2025-04-01 17:46:07 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) 2025-04-01 17:46:07 2025-04-01 17:46:07 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) 2025-04-01 17:46:07 2025-04-01 17:46:07 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) 2025-04-01 17:46:07 2025-04-01 17:46:07 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) 2025-04-01 17:46:07 2025-04-01 17:46:07 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) 2025-04-01 17:46:07 2025-04-01 17:46:07 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) 2025-04-01 17:46:07 2025-04-01 17:46:07 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) 2025-04-01 17:46:07 2025-04-01 17:46:07 , result: 2025-04-01 17:46:07 2025-04-01 17:46:07 -- disable optimize_distinct_in_order 2025-04-01 17:46:07 -- distinct all primary key columns -> ordinary distinct 2025-04-01 17:46:07 DistinctTransform 2025-04-01 17:46:07 DistinctTransform 2025-04-01 17:46:07 -- enable optimize_distinct_in_order 2025-04-01 17:46:07 -- distinct with all primary key columns -> pre-distinct optimization only 2025-04-01 17:46:07 DistinctTransform 2025-04-01 17:46:07 DistinctSortedStreamTransform 2025-04-01 17:46:07 -- distinct with primary key prefix -> pre-distinct optimization only 2025-04-01 17:46:07 DistinctTransform 2025-04-01 17:46:07 DistinctSortedStreamTransform 2025-04-01 17:46:07 -- distinct with primary key prefix and order by column in distinct -> pre-distinct and final distinct optimization 2025-04-01 17:46:07 DistinctSortedTransform 2025-04-01 17:46:07 DistinctSortedStreamTransform 2025-04-01 17:46:07 -- distinct with primary key prefix and order by the same columns -> pre-distinct and final distinct optimization 2025-04-01 17:46:07 DistinctSortedStreamTransform 2025-04-01 17:46:07 DistinctSortedStreamTransform 2025-04-01 17:46:07 -- distinct with primary key prefix and order by columns are prefix of distinct columns -> pre-distinct and final distinct optimization 2025-04-01 17:46:07 DistinctSortedTransform 2025-04-01 17:46:07 DistinctSortedStreamTransform 2025-04-01 17:46:07 -- distinct with primary key prefix and order by column in distinct but non-primary key prefix -> pre-distinct and final distinct optimization 2025-04-01 17:46:07 DistinctSortedTransform 2025-04-01 17:46:07 DistinctSortedStreamTransform 2025-04-01 17:46:07 -- distinct with primary key prefix and order by column _not_ in distinct -> pre-distinct optimization only 2025-04-01 17:46:07 DistinctTransform 2025-04-01 17:46:07 DistinctSortedStreamTransform 2025-04-01 17:46:07 -- distinct with non-primary key prefix -> ordinary distinct 2025-04-01 17:46:07 DistinctTransform 2025-04-01 17:46:07 DistinctTransform 2025-04-01 17:46:07 -- distinct with non-primary key prefix and order by column in distinct -> final distinct optimization only 2025-04-01 17:46:07 DistinctSortedTransform 2025-04-01 17:46:07 DistinctTransform 2025-04-01 17:46:07 -- distinct with non-primary key prefix and order by column _not_ in distinct -> ordinary distinct 2025-04-01 17:46:07 DistinctTransform 2025-04-01 17:46:07 DistinctTransform 2025-04-01 17:46:07 -- distinct with non-primary key prefix and order by _const_ column in distinct -> ordinary distinct 2025-04-01 17:46:07 DistinctTransform 2025-04-01 17:46:07 DistinctTransform 2025-04-01 17:46:07 -- Check reading in order for distinct 2025-04-01 17:46:07 -- disabled, distinct columns match sorting key 2025-04-01 17:46:07 algorithm: Thread 2025-04-01 17:46:07 -- enabled, distinct columns match sorting key 2025-04-01 17:46:07 algorithm: InOrder 2025-04-01 17:46:07 -- enabled, distinct columns form prefix of sorting key 2025-04-01 17:46:07 algorithm: InOrder 2025-04-01 17:46:07 -- enabled, distinct columns DON't form prefix of sorting key 2025-04-01 17:46:07 algorithm: Thread 2025-04-01 17:46:07 -- enabled, distinct columns contains constant columns, non-const columns form prefix of sorting key 2025-04-01 17:46:07 algorithm: InOrder 2025-04-01 17:46:07 -- enabled, distinct columns contains constant columns, non-const columns match prefix of sorting key 2025-04-01 17:46:07 algorithm: InOrder 2025-04-01 17:46:07 -- enabled, only part of distinct columns form prefix of sorting key 2025-04-01 17:46:07 algorithm: InOrder 2025-04-01 17:46:07 === disable new analyzer === 2025-04-01 17:46:07 -- enabled, check that sorting properties are propagated from ReadFromMergeTree till preliminary distinct 2025-04-01 17:46:07 Sorting: a ASC, b ASC 2025-04-01 17:46:07 Sorting: a ASC, b ASC 2025-04-01 17:46:07 -- check that reading in order optimization for ORDER BY and DISTINCT applied correctly in the same query 2025-04-01 17:46:07 -- disabled, check that sorting description for ReadFromMergeTree match ORDER BY columns 2025-04-01 17:46:07 Sorting: a ASC 2025-04-01 17:46:07 Sorting: a ASC 2025-04-01 17:46:07 -- enabled, check that ReadFromMergeTree sorting description is overwritten by DISTINCT optimization i.e. it contains columns from DISTINCT clause 2025-04-01 17:46:07 -- enabled, check that ReadFromMergeTree sorting description is overwritten by DISTINCT optimization, but direction used from ORDER BY clause 2025-04-01 17:46:07 -- enabled, check that ReadFromMergeTree sorting description is NOT overwritten by DISTINCT optimization (1), - it contains columns from ORDER BY clause 2025-04-01 17:46:07 -- enabled, check that ReadFromMergeTree sorting description is NOT overwritten by DISTINCT optimization (2), - direction used from ORDER BY clause 2025-04-01 17:46:07 -- enabled, check that disabling other 'read in order' optimizations do not disable distinct in order optimization 2025-04-01 17:46:07 === enable new analyzer === 2025-04-01 17:46:07 -- enabled, check that sorting properties are propagated from ReadFromMergeTree till preliminary distinct 2025-04-01 17:46:07 -- disabled, check that sorting description for ReadFromMergeTree match ORDER BY columns 2025-04-01 17:46:07 -- enabled, check that ReadFromMergeTree sorting description is overwritten by DISTINCT optimization i.e. it contains columns from DISTINCT clause 2025-04-01 17:46:07 -- enabled, check that ReadFromMergeTree sorting description is overwritten by DISTINCT optimization, but direction used from ORDER BY clause 2025-04-01 17:46:07 -- enabled, check that ReadFromMergeTree sorting description is NOT overwritten by DISTINCT optimization (1), - it contains columns from ORDER BY clause 2025-04-01 17:46:07 -- enabled, check that ReadFromMergeTree sorting description is NOT overwritten by DISTINCT optimization (2), - direction used from ORDER BY clause 2025-04-01 17:46:07 -- enabled, check that disabling other 'read in order' optimizations do not disable distinct in order optimization 2025-04-01 17:46:07 2025-04-01 17:46:07 stdout: 2025-04-01 17:46:07 -- disable optimize_distinct_in_order 2025-04-01 17:46:07 -- distinct all primary key columns -> ordinary distinct 2025-04-01 17:46:07 DistinctTransform 2025-04-01 17:46:07 DistinctTransform 2025-04-01 17:46:07 -- enable optimize_distinct_in_order 2025-04-01 17:46:07 -- distinct with all primary key columns -> pre-distinct optimization only 2025-04-01 17:46:07 DistinctTransform 2025-04-01 17:46:07 DistinctSortedStreamTransform 2025-04-01 17:46:07 -- distinct with primary key prefix -> pre-distinct optimization only 2025-04-01 17:46:07 DistinctTransform 2025-04-01 17:46:07 DistinctSortedStreamTransform 2025-04-01 17:46:07 -- distinct with primary key prefix and order by column in distinct -> pre-distinct and final distinct optimization 2025-04-01 17:46:07 DistinctSortedTransform 2025-04-01 17:46:07 DistinctSortedStreamTransform 2025-04-01 17:46:07 -- distinct with primary key prefix and order by the same columns -> pre-distinct and final distinct optimization 2025-04-01 17:46:07 DistinctSortedStreamTransform 2025-04-01 17:46:07 DistinctSortedStreamTransform 2025-04-01 17:46:07 -- distinct with primary key prefix and order by columns are prefix of distinct columns -> pre-distinct and final distinct optimization 2025-04-01 17:46:07 DistinctSortedTransform 2025-04-01 17:46:07 DistinctSortedStreamTransform 2025-04-01 17:46:07 -- distinct with primary key prefix and order by column in distinct but non-primary key prefix -> pre-distinct and final distinct optimization 2025-04-01 17:46:07 DistinctSortedTransform 2025-04-01 17:46:07 DistinctSortedStreamTransform 2025-04-01 17:46:07 -- distinct with primary key prefix and order by column _not_ in distinct -> pre-distinct optimization only 2025-04-01 17:46:07 DistinctTransform 2025-04-01 17:46:07 DistinctSortedStreamTransform 2025-04-01 17:46:07 -- distinct with non-primary key prefix -> ordinary distinct 2025-04-01 17:46:07 DistinctTransform 2025-04-01 17:46:07 DistinctTransform 2025-04-01 17:46:07 -- distinct with non-primary key prefix and order by column in distinct -> final distinct optimization only 2025-04-01 17:46:07 DistinctSortedTransform 2025-04-01 17:46:07 DistinctTransform 2025-04-01 17:46:07 -- distinct with non-primary key prefix and order by column _not_ in distinct -> ordinary distinct 2025-04-01 17:46:07 DistinctTransform 2025-04-01 17:46:07 DistinctTransform 2025-04-01 17:46:07 -- distinct with non-primary key prefix and order by _const_ column in distinct -> ordinary distinct 2025-04-01 17:46:07 DistinctTransform 2025-04-01 17:46:07 DistinctTransform 2025-04-01 17:46:07 -- Check reading in order for distinct 2025-04-01 17:46:07 -- disabled, distinct columns match sorting key 2025-04-01 17:46:07 algorithm: Thread 2025-04-01 17:46:07 -- enabled, distinct columns match sorting key 2025-04-01 17:46:07 algorithm: InOrder 2025-04-01 17:46:07 -- enabled, distinct columns form prefix of sorting key 2025-04-01 17:46:07 algorithm: InOrder 2025-04-01 17:46:07 -- enabled, distinct columns DON't form prefix of sorting key 2025-04-01 17:46:07 algorithm: Thread 2025-04-01 17:46:07 -- enabled, distinct columns contains constant columns, non-const columns form prefix of sorting key 2025-04-01 17:46:07 algorithm: InOrder 2025-04-01 17:46:07 -- enabled, distinct columns contains constant columns, non-const columns match prefix of sorting key 2025-04-01 17:46:07 algorithm: InOrder 2025-04-01 17:46:07 -- enabled, only part of distinct columns form prefix of sorting key 2025-04-01 17:46:07 algorithm: InOrder 2025-04-01 17:46:07 === disable new analyzer === 2025-04-01 17:46:07 -- enabled, check that sorting properties are propagated from ReadFromMergeTree till preliminary distinct 2025-04-01 17:46:07 Sorting: a ASC, b ASC 2025-04-01 17:46:07 Sorting: a ASC, b ASC 2025-04-01 17:46:07 -- check that reading in order optimization for ORDER BY and DISTINCT applied correctly in the same query 2025-04-01 17:46:07 -- disabled, check that sorting description for ReadFromMergeTree match ORDER BY columns 2025-04-01 17:46:07 Sorting: a ASC 2025-04-01 17:46:07 Sorting: a ASC 2025-04-01 17:46:07 -- enabled, check that ReadFromMergeTree sorting description is overwritten by DISTINCT optimization i.e. it contains columns from DISTINCT clause 2025-04-01 17:46:07 -- enabled, check that ReadFromMergeTree sorting description is overwritten by DISTINCT optimization, but direction used from ORDER BY clause 2025-04-01 17:46:07 -- enabled, check that ReadFromMergeTree sorting description is NOT overwritten by DISTINCT optimization (1), - it contains columns from ORDER BY clause 2025-04-01 17:46:07 -- enabled, check that ReadFromMergeTree sorting description is NOT overwritten by DISTINCT optimization (2), - direction used from ORDER BY clause 2025-04-01 17:46:07 -- enabled, check that disabling other 'read in order' optimizations do not disable distinct in order optimization 2025-04-01 17:46:07 === enable new analyzer === 2025-04-01 17:46:07 -- enabled, check that sorting properties are propagated from ReadFromMergeTree till preliminary distinct 2025-04-01 17:46:07 -- disabled, check that sorting description for ReadFromMergeTree match ORDER BY columns 2025-04-01 17:46:07 -- enabled, check that ReadFromMergeTree sorting description is overwritten by DISTINCT optimization i.e. it contains columns from DISTINCT clause 2025-04-01 17:46:07 -- enabled, check that ReadFromMergeTree sorting description is overwritten by DISTINCT optimization, but direction used from ORDER BY clause 2025-04-01 17:46:07 -- enabled, check that ReadFromMergeTree sorting description is NOT overwritten by DISTINCT optimization (1), - it contains columns from ORDER BY clause 2025-04-01 17:46:07 -- enabled, check that ReadFromMergeTree sorting description is NOT overwritten by DISTINCT optimization (2), - direction used from ORDER BY clause 2025-04-01 17:46:07 -- enabled, check that disabling other 'read in order' optimizations do not disable distinct in order optimization 2025-04-01 17:46:07 2025-04-01 17:46:07 2025-04-01 17:46:07 Settings used in the test: --max_insert_threads 2 --group_by_two_level_threshold 786045 --group_by_two_level_threshold_bytes 50000000 --distributed_aggregation_memory_efficient 0 --fsync_metadata 0 --output_format_parallel_formatting 0 --input_format_parallel_parsing 1 --min_chunk_bytes_for_parallel_parsing 4698335 --max_read_buffer_size 789110 --prefer_localhost_replica 0 --max_block_size 86889 --max_joined_block_size_rows 67122 --max_threads 3 --optimize_append_index 1 --optimize_if_chain_to_multiif 0 --optimize_if_transform_strings_to_enum 0 --optimize_read_in_order 0 --optimize_or_like_chain 0 --optimize_substitute_columns 1 --enable_multiple_prewhere_read_steps 0 --read_in_order_two_level_merge_threshold 56 --optimize_aggregation_in_order 0 --aggregation_in_order_max_block_bytes 29243934 --use_uncompressed_cache 0 --min_bytes_to_use_direct_io 9709034220 --min_bytes_to_use_mmap_io 3483143077 --local_filesystem_read_method pread --remote_filesystem_read_method read --local_filesystem_read_prefetch 1 --filesystem_cache_segments_batch_size 2 --read_from_filesystem_cache_if_exists_otherwise_bypass_cache 1 --throw_on_error_from_cache_on_write_operations 0 --remote_filesystem_read_prefetch 0 --allow_prefetched_read_pool_for_remote_filesystem 1 --filesystem_prefetch_max_memory_usage 32Mi --filesystem_prefetches_limit 0 --filesystem_prefetch_min_bytes_for_single_read_task 16Mi --filesystem_prefetch_step_marks 0 --filesystem_prefetch_step_bytes 0 --compile_aggregate_expressions 0 --compile_sort_description 0 --merge_tree_coarse_index_granularity 22 --optimize_distinct_in_order 1 --max_bytes_before_external_sort 10737418240 --max_bytes_before_external_group_by 10711791914 --max_bytes_before_remerge_sort 1724396297 --min_compress_block_size 509181 --max_compress_block_size 961470 --merge_tree_compact_parts_min_granules_to_multibuffer_read 113 --optimize_sorting_by_input_stream_properties 0 --http_response_buffer_size 10285867 --http_wait_end_of_query True --enable_memory_bound_merging_of_aggregation_results 1 --min_count_to_compile_expression 0 --min_count_to_compile_aggregate_expression 0 --min_count_to_compile_sort_description 3 --session_timezone Mexico/BajaSur --use_page_cache_for_disks_without_file_cache True --page_cache_inject_eviction True --merge_tree_read_split_ranges_into_intersecting_and_non_intersecting_injection_probability 0.89 --prefer_external_sort_block_bytes 1 --cross_join_min_rows_to_compress 1 --cross_join_min_bytes_to_compress 1 --min_external_table_block_size_bytes 100000000 --max_parsing_threads 0 --optimize_functions_to_subcolumns 0 --parallel_replicas_local_plan 1 --query_plan_join_swap_table auto --output_format_native_write_json_as_string 1 --enable_vertical_final 1 2025-04-01 17:46:07 2025-04-01 17:46:07 MergeTree settings used in test: --ratio_of_defaults_for_sparse_serialization 0.7563738772706338 --prefer_fetch_merged_part_size_threshold 2012106554 --vertical_merge_algorithm_min_rows_to_activate 1 --vertical_merge_algorithm_min_columns_to_activate 82 --allow_vertical_merges_from_compact_to_wide_parts 0 --min_merge_bytes_to_use_direct_io 10737418240 --index_granularity_bytes 27699374 --merge_max_block_size 19581 --index_granularity 48176 --min_bytes_for_wide_part 0 --marks_compress_block_size 17345 --primary_key_compress_block_size 29004 --replace_long_file_name_to_hash 1 --max_file_name_length 0 --min_bytes_for_full_part_storage 0 --compact_parts_max_bytes_to_buffer 126996618 --compact_parts_max_granules_to_buffer 256 --compact_parts_merge_max_bytes_to_prefetch_part 22889841 --cache_populated_by_fetch 1 --concurrent_part_removal_threshold 100 --old_parts_lifetime 480 --prewarm_mark_cache 1 --use_const_adaptive_granularity 1 --enable_index_granularity_compression 0 --use_primary_key_cache 1 --prewarm_primary_key_cache 0 2025-04-01 17:46:07 2025-04-01 17:46:07 Database: test_l0nrtmbt 2025-04-01 17:46:13 02122_parallel_formatting_RowBinaryWithNames: [ FAIL ] 0.00 sec. 2025-04-01 17:46:13 Reason: server died 2025-04-01 17:46:13 Server does not respond to health check 2025-04-01 17:46:13 2025-04-01 17:46:13 01440_big_int_exotic_casts: [ FAIL ] 0.00 sec. 2025-04-01 17:46:13 Reason: server died 2025-04-01 17:46:13 Server does not respond to health check 2025-04-01 17:46:13 2025-04-01 17:46:13 Process Process-6: 2025-04-01 17:46:13 Traceback (most recent call last): 2025-04-01 17:46:13 File "/usr/lib/python3.10/multiprocessing/process.py", line 314, in _bootstrap 2025-04-01 17:46:13 self.run() 2025-04-01 17:46:13 File "/usr/lib/python3.10/multiprocessing/process.py", line 108, in run 2025-04-01 17:46:13 self._target(*self._args, **self._kwargs) 2025-04-01 17:46:13 File "/usr/bin/clickhouse-test", line 2626, in run_tests_process 2025-04-01 17:46:13 return run_tests_array(*args, **kwargs) 2025-04-01 17:46:13 File "/usr/bin/clickhouse-test", line 2407, in run_tests_array 2025-04-01 17:46:13 stop_tests() 2025-04-01 17:46:13 File "/usr/bin/clickhouse-test", line 452, in stop_tests 2025-04-01 17:46:13 cleanup_child_processes(os.getpid()) 2025-04-01 17:46:13 File "/usr/bin/clickhouse-test", line 440, in cleanup_child_processes 2025-04-01 17:46:13 child_pgid = os.getpgid(child) 2025-04-01 17:46:13 ProcessLookupError: [Errno 3] No such process 2025-04-01 17:46:13 Child processes of 1388: 2025-04-01 17:46:13 41795 /bin/sh -c pgrep --parent 1388 -a 2025-04-01 17:46:13 Process Process-4: 2025-04-01 17:46:13 Traceback (most recent call last): 2025-04-01 17:46:13 File "/usr/lib/python3.10/multiprocessing/process.py", line 314, in _bootstrap 2025-04-01 17:46:13 self.run() 2025-04-01 17:46:13 File "/usr/lib/python3.10/multiprocessing/process.py", line 108, in run 2025-04-01 17:46:13 self._target(*self._args, **self._kwargs) 2025-04-01 17:46:13 File "/usr/bin/clickhouse-test", line 2626, in run_tests_process 2025-04-01 17:46:13 return run_tests_array(*args, **kwargs) 2025-04-01 17:46:13 File "/usr/bin/clickhouse-test", line 2407, in run_tests_array 2025-04-01 17:46:13 stop_tests() 2025-04-01 17:46:13 File "/usr/bin/clickhouse-test", line 452, in stop_tests 2025-04-01 17:46:13 cleanup_child_processes(os.getpid()) 2025-04-01 17:46:13 File "/usr/bin/clickhouse-test", line 440, in cleanup_child_processes 2025-04-01 17:46:13 child_pgid = os.getpgid(child) 2025-04-01 17:46:13 ProcessLookupError: [Errno 3] No such process 2025-04-01 17:46:13 Child processes of 1384: 2025-04-01 17:46:13 41797 /bin/sh -c pgrep --parent 1384 -a 2025-04-01 17:46:13 02221_parallel_replicas_bug: [ FAIL ] 0.00 sec. 2025-04-01 17:46:13 Reason: server died 2025-04-01 17:46:13 Server does not respond to health check 2025-04-01 17:46:13 2025-04-01 17:46:13 Database: test_sczjo540 2025-04-01 17:46:13 Process Process-5: 2025-04-01 17:46:13 Traceback (most recent call last): 2025-04-01 17:46:13 File "/usr/lib/python3.10/multiprocessing/process.py", line 314, in _bootstrap 2025-04-01 17:46:13 self.run() 2025-04-01 17:46:13 File "/usr/lib/python3.10/multiprocessing/process.py", line 108, in run 2025-04-01 17:46:13 self._target(*self._args, **self._kwargs) 2025-04-01 17:46:13 File "/usr/bin/clickhouse-test", line 2626, in run_tests_process 2025-04-01 17:46:13 return run_tests_array(*args, **kwargs) 2025-04-01 17:46:13 File "/usr/bin/clickhouse-test", line 2407, in run_tests_array 2025-04-01 17:46:13 stop_tests() 2025-04-01 17:46:13 File "/usr/bin/clickhouse-test", line 452, in stop_tests 2025-04-01 17:46:13 cleanup_child_processes(os.getpid()) 2025-04-01 17:46:13 File "/usr/bin/clickhouse-test", line 440, in cleanup_child_processes 2025-04-01 17:46:13 child_pgid = os.getpgid(child) 2025-04-01 17:46:13 ProcessLookupError: [Errno 3] No such process 2025-04-01 17:46:13 Child processes of 1385: 2025-04-01 17:46:13 41807 /bin/sh -c pgrep --parent 1385 -a 2025-04-01 17:46:14 01901_in_literal_shard_prune: [ FAIL ] 0.00 sec. 2025-04-01 17:46:14 Reason: server died 2025-04-01 17:46:14 Server does not respond to health check 2025-04-01 17:46:14 2025-04-01 17:46:14 Process Process-8: 2025-04-01 17:46:14 Traceback (most recent call last): 2025-04-01 17:46:14 File "/usr/lib/python3.10/multiprocessing/process.py", line 314, in _bootstrap 2025-04-01 17:46:14 self.run() 2025-04-01 17:46:14 File "/usr/lib/python3.10/multiprocessing/process.py", line 108, in run 2025-04-01 17:46:14 self._target(*self._args, **self._kwargs) 2025-04-01 17:46:14 File "/usr/bin/clickhouse-test", line 2626, in run_tests_process 2025-04-01 17:46:14 return run_tests_array(*args, **kwargs) 2025-04-01 17:46:14 File "/usr/bin/clickhouse-test", line 2407, in run_tests_array 2025-04-01 17:46:14 stop_tests() 2025-04-01 17:46:14 File "/usr/bin/clickhouse-test", line 452, in stop_tests 2025-04-01 17:46:14 cleanup_child_processes(os.getpid()) 2025-04-01 17:46:14 File "/usr/bin/clickhouse-test", line 440, in cleanup_child_processes 2025-04-01 17:46:14 child_pgid = os.getpgid(child) 2025-04-01 17:46:14 ProcessLookupError: [Errno 3] No such process 2025-04-01 17:46:14 Child processes of 1391: 2025-04-01 17:46:14 41813 /bin/sh -c pgrep --parent 1391 -a 2025-04-01 17:46:14 02565_update_empty_nested: [ FAIL ] 0.00 sec. 2025-04-01 17:46:14 Reason: server died 2025-04-01 17:46:14 Server does not respond to health check 2025-04-01 17:46:14 2025-04-01 17:46:14 Process Process-7: 2025-04-01 17:46:14 Traceback (most recent call last): 2025-04-01 17:46:14 File "/usr/lib/python3.10/multiprocessing/process.py", line 314, in _bootstrap 2025-04-01 17:46:14 self.run() 2025-04-01 17:46:14 File "/usr/lib/python3.10/multiprocessing/process.py", line 108, in run 2025-04-01 17:46:14 self._target(*self._args, **self._kwargs) 2025-04-01 17:46:14 File "/usr/bin/clickhouse-test", line 2626, in run_tests_process 2025-04-01 17:46:14 return run_tests_array(*args, **kwargs) 2025-04-01 17:46:14 File "/usr/bin/clickhouse-test", line 2407, in run_tests_array 2025-04-01 17:46:14 stop_tests() 2025-04-01 17:46:14 File "/usr/bin/clickhouse-test", line 452, in stop_tests 2025-04-01 17:46:14 cleanup_child_processes(os.getpid()) 2025-04-01 17:46:14 File "/usr/bin/clickhouse-test", line 440, in cleanup_child_processes 2025-04-01 17:46:14 child_pgid = os.getpgid(child) 2025-04-01 17:46:14 ProcessLookupError: [Errno 3] No such process 2025-04-01 17:46:14 Child processes of 1390: 2025-04-01 17:46:14 41819 /bin/sh -c pgrep --parent 1390 -a 2025-04-01 17:46:15 01115_prewhere_array_join: [ FAIL ] 0.00 sec. 2025-04-01 17:46:15 Reason: server died 2025-04-01 17:46:15 Server does not respond to health check 2025-04-01 17:46:15 2025-04-01 17:46:15 Process Process-3: 2025-04-01 17:46:15 Traceback (most recent call last): 2025-04-01 17:46:15 File "/usr/lib/python3.10/multiprocessing/process.py", line 314, in _bootstrap 2025-04-01 17:46:15 self.run() 2025-04-01 17:46:15 File "/usr/lib/python3.10/multiprocessing/process.py", line 108, in run 2025-04-01 17:46:15 self._target(*self._args, **self._kwargs) 2025-04-01 17:46:15 File "/usr/bin/clickhouse-test", line 2626, in run_tests_process 2025-04-01 17:46:15 return run_tests_array(*args, **kwargs) 2025-04-01 17:46:15 File "/usr/bin/clickhouse-test", line 2407, in run_tests_array 2025-04-01 17:46:15 stop_tests() 2025-04-01 17:46:15 File "/usr/bin/clickhouse-test", line 452, in stop_tests 2025-04-01 17:46:15 cleanup_child_processes(os.getpid()) 2025-04-01 17:46:15 File "/usr/bin/clickhouse-test", line 440, in cleanup_child_processes 2025-04-01 17:46:15 child_pgid = os.getpgid(child) 2025-04-01 17:46:15 ProcessLookupError: [Errno 3] No such process 2025-04-01 17:46:15 Child processes of 1383: 2025-04-01 17:46:15 41825 /bin/sh -c pgrep --parent 1383 -a 2025-04-01 17:46:17 01047_no_alias_columns_with_table_aliases: [ FAIL ] 0.00 sec. 2025-04-01 17:46:17 Reason: server died 2025-04-01 17:46:17 Server does not respond to health check 2025-04-01 17:46:17 2025-04-01 17:46:17 Process Process-10: 2025-04-01 17:46:17 Traceback (most recent call last): 2025-04-01 17:46:17 File "/usr/lib/python3.10/multiprocessing/process.py", line 314, in _bootstrap 2025-04-01 17:46:17 self.run() 2025-04-01 17:46:17 File "/usr/lib/python3.10/multiprocessing/process.py", line 108, in run 2025-04-01 17:46:17 self._target(*self._args, **self._kwargs) 2025-04-01 17:46:17 File "/usr/bin/clickhouse-test", line 2626, in run_tests_process 2025-04-01 17:46:17 return run_tests_array(*args, **kwargs) 2025-04-01 17:46:17 File "/usr/bin/clickhouse-test", line 2407, in run_tests_array 2025-04-01 17:46:17 stop_tests() 2025-04-01 17:46:17 File "/usr/bin/clickhouse-test", line 452, in stop_tests 2025-04-01 17:46:17 cleanup_child_processes(os.getpid()) 2025-04-01 17:46:17 File "/usr/bin/clickhouse-test", line 440, in cleanup_child_processes 2025-04-01 17:46:17 child_pgid = os.getpgid(child) 2025-04-01 17:46:17 ProcessLookupError: [Errno 3] No such process 2025-04-01 17:46:17 Child processes of 1397: 2025-04-01 17:46:17 41831 /bin/sh -c pgrep --parent 1397 -a 2025-04-01 17:59:04 Hung check failed: [Errno 111] Connection refused 2025-04-01 17:59:04 Server died, terminating all processes... 2025-04-01 17:59:04 Running 130 stateless tests (MainProcess). 2025-04-01 17:59:04 Child processes of 1367: 2025-04-01 17:59:04 1370 python3 /usr/bin/clickhouse-test --testname --shard --zookeeper --check-zookeeper-session --hung-check --print-time --no-drop-if-fail --capture-client-stacktrace --queries /repo/tests/queries --test-runs 1 --hung-check --print-time --jobs 8 --run-by-hash-num 2 --run-by-hash-total 4 --report-logs-stats 2025-04-01 17:59:04 1376 python3 /usr/bin/clickhouse-test --testname --shard --zookeeper --check-zookeeper-session --hung-check --print-time --no-drop-if-fail --capture-client-stacktrace --queries /repo/tests/queries --test-runs 1 --hung-check --print-time --jobs 8 --run-by-hash-num 2 --run-by-hash-total 4 --report-logs-stats 2025-04-01 17:59:04 41839 /bin/sh -c pgrep --parent 1367 -a 2025-04-01 17:59:04 Traceback (most recent call last): 2025-04-01 17:59:04 File "/usr/bin/clickhouse-test", line 3716, in 2025-04-01 17:59:04 main(args) 2025-04-01 17:59:04 File "/usr/bin/clickhouse-test", line 3063, in main 2025-04-01 17:59:04 total_tests_run += do_run_tests(args.jobs, test_suite) 2025-04-01 17:59:04 File "/usr/bin/clickhouse-test", line 2697, in do_run_tests 2025-04-01 17:59:04 run_tests_array( 2025-04-01 17:59:04 File "/usr/bin/clickhouse-test", line 2343, in run_tests_array 2025-04-01 17:59:04 stop_tests() 2025-04-01 17:59:04 File "/usr/bin/clickhouse-test", line 452, in stop_tests 2025-04-01 17:59:04 cleanup_child_processes(os.getpid()) 2025-04-01 17:59:04 File "/usr/bin/clickhouse-test", line 440, in cleanup_child_processes 2025-04-01 17:59:04 child_pgid = os.getpgid(child) 2025-04-01 17:59:04 ProcessLookupError: [Errno 3] No such process + set -e + echo 'Files in current directory' + ls -la ./ Files in current directory total 127172 drwxr-xr-x 1 root root 4096 Apr 1 17:45 . drwxr-xr-x 1 root root 4096 Apr 1 17:45 .. lrwxrwxrwx 1 root root 7 Sep 11 2024 bin -> usr/bin drwxr-xr-x 2 root root 4096 Apr 18 2022 boot drwxr-xr-x 14 root root 3840 Apr 1 17:39 dev -rwxr-xr-x 1 root root 0 Apr 1 17:39 .dockerenv drwxr-xr-x 1 root root 4096 Apr 1 17:39 etc drwxr-xr-x 2 root root 4096 Apr 18 2022 home lrwxrwxrwx 1 root root 7 Sep 11 2024 lib -> usr/lib lrwxrwxrwx 1 root root 9 Sep 11 2024 lib32 -> usr/lib32 lrwxrwxrwx 1 root root 9 Sep 11 2024 lib64 -> usr/lib64 lrwxrwxrwx 1 root root 10 Sep 11 2024 libx32 -> usr/libx32 -rwxr-xr-x 1 root root 26927256 Jan 15 17:53 mc drwxr-xr-x 2 root root 4096 Sep 11 2024 media -rwxr-xr-x 1 root root 103174296 Jan 15 17:53 minio drwxr-xr-x 4 root root 4096 Apr 1 17:39 minio_data drwxr-xr-x 2 root root 4096 Sep 11 2024 mnt drwxr-xr-x 1 root root 4096 Jan 14 17:08 opt -rw-r--r-- 1 root root 0 Feb 14 2024 .package-cache-mutate drwxrwxr-x 2 1000 1000 4096 Apr 1 17:39 package_folder drwxr-x--- 2 root root 4096 Apr 1 17:41 preprocessed_configs dr-xr-xr-x 305 root root 0 Apr 1 17:39 proc -rw-r--r-- 1 root root 29 Apr 1 17:40 queries_02352 drwxrwxr-x 17 1000 1000 4096 Apr 1 17:36 repo -rw-rw-r-- 1 root root 863 Jan 15 17:27 requirements.txt drwx------ 1 root root 4096 Apr 1 17:45 root drwxr-xr-x 1 root root 4096 Apr 1 17:39 run lrwxrwxrwx 1 root root 8 Sep 11 2024 sbin -> usr/sbin -rw-r--r-- 1 root root 747 Apr 1 17:39 script.gdb -rwxrwxr-x 1 root root 10851 Jan 14 17:07 setup_export_logs.sh drwxr-xr-x 2 root root 4096 Sep 11 2024 srv dr-xr-xr-x 13 root root 0 Apr 1 17:39 sys drwxrwxr-x 2 1000 1000 4096 Apr 1 17:39 test_output drwxrwxrwt 1 root root 4096 Apr 1 17:59 tmp drwxr-xr-x 1 root root 4096 Sep 11 2024 usr drwxr-xr-x 1 root root 4096 Sep 11 2024 var + echo 'Files in root directory' + ls -la / Files in root directory total 127172 drwxr-xr-x 1 root root 4096 Apr 1 17:45 . drwxr-xr-x 1 root root 4096 Apr 1 17:45 .. lrwxrwxrwx 1 root root 7 Sep 11 2024 bin -> usr/bin drwxr-xr-x 2 root root 4096 Apr 18 2022 boot drwxr-xr-x 14 root root 3840 Apr 1 17:39 dev -rwxr-xr-x 1 root root 0 Apr 1 17:39 .dockerenv drwxr-xr-x 1 root root 4096 Apr 1 17:39 etc drwxr-xr-x 2 root root 4096 Apr 18 2022 home lrwxrwxrwx 1 root root 7 Sep 11 2024 lib -> usr/lib lrwxrwxrwx 1 root root 9 Sep 11 2024 lib32 -> usr/lib32 lrwxrwxrwx 1 root root 9 Sep 11 2024 lib64 -> usr/lib64 lrwxrwxrwx 1 root root 10 Sep 11 2024 libx32 -> usr/libx32 -rwxr-xr-x 1 root root 26927256 Jan 15 17:53 mc drwxr-xr-x 2 root root 4096 Sep 11 2024 media -rwxr-xr-x 1 root root 103174296 Jan 15 17:53 minio drwxr-xr-x 4 root root 4096 Apr 1 17:39 minio_data drwxr-xr-x 2 root root 4096 Sep 11 2024 mnt drwxr-xr-x 1 root root 4096 Jan 14 17:08 opt -rw-r--r-- 1 root root 0 Feb 14 2024 .package-cache-mutate drwxrwxr-x 2 1000 1000 4096 Apr 1 17:39 package_folder drwxr-x--- 2 root root 4096 Apr 1 17:41 preprocessed_configs dr-xr-xr-x 305 root root 0 Apr 1 17:39 proc -rw-r--r-- 1 root root 29 Apr 1 17:40 queries_02352 drwxrwxr-x 17 1000 1000 4096 Apr 1 17:36 repo -rw-rw-r-- 1 root root 863 Jan 15 17:27 requirements.txt drwx------ 1 root root 4096 Apr 1 17:45 root drwxr-xr-x 1 root root 4096 Apr 1 17:39 run lrwxrwxrwx 1 root root 8 Sep 11 2024 sbin -> usr/sbin -rw-r--r-- 1 root root 747 Apr 1 17:39 script.gdb -rwxrwxr-x 1 root root 10851 Jan 14 17:07 setup_export_logs.sh drwxr-xr-x 2 root root 4096 Sep 11 2024 srv dr-xr-xr-x 13 root root 0 Apr 1 17:39 sys drwxrwxr-x 2 1000 1000 4096 Apr 1 17:39 test_output drwxrwxrwt 1 root root 4096 Apr 1 17:59 tmp drwxr-xr-x 1 root root 4096 Sep 11 2024 usr drwxr-xr-x 1 root root 4096 Sep 11 2024 var + clickhouse-client -q 'system flush logs' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + : + stop_logs_replication + echo 'Detach all logs replication' Detach all logs replication + clickhouse-client --query 'select database||'\''.'\''||table from system.tables where database = '\''system'\'' and (table like '\''%_sender'\'' or table like '\''%_watcher'\'')' + tee /dev/stderr + timeout --preserve-status --signal TERM --kill-after 5m 15m xargs -n1 -r -i clickhouse-client --query 'drop table {}' xargs: warning: options --max-args and --replace/-I/-i are mutually exclusive, ignoring previous --max-args value Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + logs_saver_client_options='--max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0' + failed_to_save_logs=0 + for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log + clickhouse-client --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 -q 'select * from system.query_log into outfile '\''/test_output/query_log.tsv.zst'\'' format TSVWithNamesAndTypes' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + failed_to_save_logs=1 + [[ 0 -eq 1 ]] + [[ 0 -eq 1 ]] + for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log + clickhouse-client --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 -q 'select * from system.zookeeper_log into outfile '\''/test_output/zookeeper_log.tsv.zst'\'' format TSVWithNamesAndTypes' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + failed_to_save_logs=1 + [[ 0 -eq 1 ]] + [[ 0 -eq 1 ]] + for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log + clickhouse-client --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 -q 'select * from system.trace_log into outfile '\''/test_output/trace_log.tsv.zst'\'' format TSVWithNamesAndTypes' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + failed_to_save_logs=1 + [[ 0 -eq 1 ]] + [[ 0 -eq 1 ]] + for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log + clickhouse-client --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 -q 'select * from system.transactions_info_log into outfile '\''/test_output/transactions_info_log.tsv.zst'\'' format TSVWithNamesAndTypes' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + failed_to_save_logs=1 + [[ 0 -eq 1 ]] + [[ 0 -eq 1 ]] + for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log + clickhouse-client --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 -q 'select * from system.metric_log into outfile '\''/test_output/metric_log.tsv.zst'\'' format TSVWithNamesAndTypes' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + failed_to_save_logs=1 + [[ 0 -eq 1 ]] + [[ 0 -eq 1 ]] + for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log + clickhouse-client --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 -q 'select * from system.blob_storage_log into outfile '\''/test_output/blob_storage_log.tsv.zst'\'' format TSVWithNamesAndTypes' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + failed_to_save_logs=1 + [[ 0 -eq 1 ]] + [[ 0 -eq 1 ]] + for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log + clickhouse-client --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 -q 'select * from system.error_log into outfile '\''/test_output/error_log.tsv.zst'\'' format TSVWithNamesAndTypes' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + failed_to_save_logs=1 + [[ 0 -eq 1 ]] + [[ 0 -eq 1 ]] + for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log + clickhouse-client --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 -q 'select * from system.query_metric_log into outfile '\''/test_output/query_metric_log.tsv.zst'\'' format TSVWithNamesAndTypes' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + failed_to_save_logs=1 + [[ 0 -eq 1 ]] + [[ 0 -eq 1 ]] + sleep 1 + clickhouse-client -q 'SYSTEM FLUSH ASYNC INSERT QUEUE' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + : + clickhouse-client --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 -q 'SELECT log FROM minio_audit_logs ORDER BY log.time INTO OUTFILE '\''/test_output/minio_audit_logs.jsonl.zst'\'' FORMAT JSONEachRow' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + : + clickhouse-client --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 -q 'SELECT log FROM minio_server_logs ORDER BY log.time INTO OUTFILE '\''/test_output/minio_server_logs.jsonl.zst'\'' FORMAT JSONEachRow' Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) + : + sudo clickhouse stop ==41896==Installed the sigaction for signal 11 ==41896==Installed the sigaction for signal 7 ==41896==Installed the sigaction for signal 8 ==41896==Using llvm-symbolizer found at: /usr/bin/llvm-symbolizer-19 ***** Running under ThreadSanitizer v3 (pid 41896) ***** ThreadSanitizer: growing sync allocator: 0 out of 1048576*1024 ThreadSanitizer: growing heap block allocator: 0 out of 262144*4096 ThreadSanitizer: growing heap block allocator: 1 out of 262144*4096 ThreadSanitizer: growing sync allocator: 1 out of 1048576*1024 ThreadSanitizer: growing heap block allocator: 2 out of 262144*4096 ThreadSanitizer: growing sync allocator: 2 out of 1048576*1024 ThreadSanitizer: growing sync allocator: 3 out of 1048576*1024 ThreadSanitizer: growing sync allocator: 4 out of 1048576*1024 ThreadSanitizer: growing heap block allocator: 3 out of 262144*4096 ThreadSanitizer: growing sync allocator: 5 out of 1048576*1024 ThreadSanitizer: growing sync allocator: 6 out of 1048576*1024 /var/run/clickhouse-server/clickhouse-server.pid file exists and contains pid = 428. The process with pid = 428 does not exist. Stats: SizeClassAllocator64: 10M mapped (1M rss) in 20235 allocations; remains 5797 01 ( 16): mapped: 256K allocs: 1024 frees: 768 inuse: 256 num_freed_chunks 16128 avail: 16384 rss: 16K releases: 0 last released: 0K region: 0x720400000000 02 ( 32): mapped: 256K allocs: 7168 frees: 6656 inuse: 512 num_freed_chunks 7680 avail: 8192 rss: 176K releases: 0 last released: 0K region: 0x720800000000 03 ( 48): mapped: 256K allocs: 5632 frees: 4608 inuse: 1024 num_freed_chunks 4437 avail: 5461 rss: 252K releases: 0 last released: 0K region: 0x720c00000000 04 ( 64): mapped: 256K allocs: 1792 frees: 1408 inuse: 384 num_freed_chunks 3712 avail: 4096 rss: 108K releases: 0 last released: 0K region: 0x721000000000 05 ( 80): mapped: 256K allocs: 384 frees: 128 inuse: 256 num_freed_chunks 3020 avail: 3276 rss: 28K releases: 0 last released: 0K region: 0x721400000000 06 ( 96): mapped: 256K allocs: 256 frees: 0 inuse: 256 num_freed_chunks 2474 avail: 2730 rss: 20K releases: 0 last released: 0K region: 0x721800000000 07 ( 112): mapped: 256K allocs: 384 frees: 128 inuse: 256 num_freed_chunks 2084 avail: 2340 rss: 36K releases: 0 last released: 0K region: 0x721c00000000 08 ( 128): mapped: 256K allocs: 640 frees: 384 inuse: 256 num_freed_chunks 1792 avail: 2048 rss: 80K releases: 0 last released: 0K region: 0x722000000000 09 ( 144): mapped: 256K allocs: 128 frees: 0 inuse: 128 num_freed_chunks 1692 avail: 1820 rss: 8K releases: 0 last released: 0K region: 0x722400000000 10 ( 160): mapped: 256K allocs: 128 frees: 0 inuse: 128 num_freed_chunks 1510 avail: 1638 rss: 8K releases: 0 last released: 0K region: 0x722800000000 11 ( 176): mapped: 256K allocs: 128 frees: 0 inuse: 128 num_freed_chunks 1361 avail: 1489 rss: 4K releases: 0 last released: 0K region: 0x722c00000000 12 ( 192): mapped: 256K allocs: 128 frees: 0 inuse: 128 num_freed_chunks 1237 avail: 1365 rss: 16K releases: 0 last released: 0K region: 0x723000000000 13 ( 208): mapped: 256K allocs: 128 frees: 0 inuse: 128 num_freed_chunks 1132 avail: 1260 rss: 4K releases: 0 last released: 0K region: 0x723400000000 14 ( 224): mapped: 256K allocs: 512 frees: 256 inuse: 256 num_freed_chunks 914 avail: 1170 rss: 92K releases: 0 last released: 0K region: 0x723800000000 15 ( 240): mapped: 256K allocs: 256 frees: 0 inuse: 256 num_freed_chunks 836 avail: 1092 rss: 40K releases: 0 last released: 0K region: 0x723c00000000 16 ( 256): mapped: 256K allocs: 256 frees: 0 inuse: 256 num_freed_chunks 768 avail: 1024 rss: 40K releases: 0 last released: 0K region: 0x724000000000 17 ( 320): mapped: 256K allocs: 128 frees: 0 inuse: 128 num_freed_chunks 691 avail: 819 rss: 16K releases: 0 last released: 0K region: 0x724400000000 18 ( 384): mapped: 256K allocs: 128 frees: 0 inuse: 128 num_freed_chunks 554 avail: 682 rss: 8K releases: 0 last released: 0K region: 0x724800000000 19 ( 448): mapped: 256K allocs: 128 frees: 0 inuse: 128 num_freed_chunks 457 avail: 585 rss: 8K releases: 0 last released: 0K region: 0x724c00000000 20 ( 512): mapped: 256K allocs: 128 frees: 0 inuse: 128 num_freed_chunks 384 avail: 512 rss: 4K releases: 0 last released: 0K region: 0x725000000000 21 ( 640): mapped: 256K allocs: 102 frees: 0 inuse: 102 num_freed_chunks 307 avail: 409 rss: 4K releases: 0 last released: 0K region: 0x725400000000 22 ( 768): mapped: 256K allocs: 85 frees: 0 inuse: 85 num_freed_chunks 256 avail: 341 rss: 4K releases: 0 last released: 0K region: 0x725800000000 23 ( 896): mapped: 256K allocs: 73 frees: 0 inuse: 73 num_freed_chunks 219 avail: 292 rss: 8K releases: 0 last released: 0K region: 0x725c00000000 24 ( 1024): mapped: 256K allocs: 64 frees: 0 inuse: 64 num_freed_chunks 192 avail: 256 rss: 8K releases: 0 last released: 0K region: 0x726000000000 25 ( 1280): mapped: 256K allocs: 204 frees: 102 inuse: 102 num_freed_chunks 102 avail: 204 rss: 252K releases: 0 last released: 0K region: 0x726400000000 26 ( 1536): mapped: 256K allocs: 42 frees: 0 inuse: 42 num_freed_chunks 128 avail: 170 rss: 4K releases: 0 last released: 0K region: 0x726800000000 27 ( 1792): mapped: 256K allocs: 36 frees: 0 inuse: 36 num_freed_chunks 110 avail: 146 rss: 4K releases: 0 last released: 0K region: 0x726c00000000 28 ( 2048): mapped: 256K allocs: 32 frees: 0 inuse: 32 num_freed_chunks 96 avail: 128 rss: 8K releases: 0 last released: 0K region: 0x727000000000 29 ( 2560): mapped: 256K allocs: 25 frees: 0 inuse: 25 num_freed_chunks 77 avail: 102 rss: 20K releases: 0 last released: 0K region: 0x727400000000 30 ( 3072): mapped: 256K allocs: 21 frees: 0 inuse: 21 num_freed_chunks 64 avail: 85 rss: 8K releases: 0 last released: 0K region: 0x727800000000 31 ( 3584): mapped: 256K allocs: 18 frees: 0 inuse: 18 num_freed_chunks 55 avail: 73 rss: 12K releases: 0 last released: 0K region: 0x727c00000000 32 ( 4096): mapped: 256K allocs: 16 frees: 0 inuse: 16 num_freed_chunks 48 avail: 64 rss: 12K releases: 0 last released: 0K region: 0x728000000000 33 ( 5120): mapped: 256K allocs: 12 frees: 0 inuse: 12 num_freed_chunks 39 avail: 51 rss: 16K releases: 0 last released: 0K region: 0x728400000000 34 ( 6144): mapped: 256K allocs: 10 frees: 0 inuse: 10 num_freed_chunks 32 avail: 42 rss: 12K releases: 0 last released: 0K region: 0x728800000000 35 ( 7168): mapped: 256K allocs: 9 frees: 0 inuse: 9 num_freed_chunks 27 avail: 36 rss: 52K releases: 0 last released: 0K region: 0x728c00000000 36 ( 8192): mapped: 256K allocs: 8 frees: 0 inuse: 8 num_freed_chunks 24 avail: 32 rss: 16K releases: 0 last released: 0K region: 0x729000000000 37 ( 10240): mapped: 256K allocs: 6 frees: 0 inuse: 6 num_freed_chunks 19 avail: 25 rss: 12K releases: 0 last released: 0K region: 0x729400000000 38 ( 12288): mapped: 256K allocs: 5 frees: 0 inuse: 5 num_freed_chunks 16 avail: 21 rss: 12K releases: 0 last released: 0K region: 0x729800000000 39 ( 14336): mapped: 256K allocs: 4 frees: 0 inuse: 4 num_freed_chunks 14 avail: 18 rss: 16K releases: 0 last released: 0K region: 0x729c00000000 42 ( 24576): mapped: 256K allocs: 2 frees: 0 inuse: 2 num_freed_chunks 8 avail: 10 rss: 16K releases: 0 last released: 0K region: 0x72a800000000 43 ( 28672): mapped: 256K allocs: 2 frees: 0 inuse: 2 num_freed_chunks 7 avail: 9 rss: 28K releases: 0 last released: 0K region: 0x72ac00000000 44 ( 32768): mapped: 256K allocs: 2 frees: 0 inuse: 2 num_freed_chunks 6 avail: 8 rss: 32K releases: 0 last released: 0K region: 0x72b000000000 45 ( 40960): mapped: 256K allocs: 1 frees: 0 inuse: 1 num_freed_chunks 5 avail: 6 rss: 4K releases: 0 last released: 0K region: 0x72b400000000 Stats: LargeMmapAllocator: allocated 1 times, remains 0 (0 K) max 1 M; by size logs: 20:1; + [[ 0 -eq 1 ]] + [[ 0 -eq 1 ]] + kill 1344 + rg -Fa '' /var/log/clickhouse-server/clickhouse-server.log 2025.04.01 17:45:29.274291 [ 2230 ] {549c2d11-5172-44cf-9edd-d2571bb9f6a9} : Logical error: 'Replica info is not initialized'. 2025.04.01 17:45:29.375723 [ 2230 ] {549c2d11-5172-44cf-9edd-d2571bb9f6a9} : Stack trace (when copying this message, always include the lines below): 2025.04.01 17:45:29.377757 [ 40158 ] {} BaseDaemon: ########## Short fault info ############ 2025.04.01 17:45:29.378085 [ 40158 ] {} BaseDaemon: (version 24.12.2.20221.altinityantalya (altinity build), build id: 1D546DEA3F7DAB479A2081B78470431AC256AC96, git hash: 82252d159dc02cab0f366aaa5691adc1545dd11d, architecture: x86_64) (from thread 2230) Received signal 6 2025.04.01 17:45:29.378470 [ 40158 ] {} BaseDaemon: Signal description: Aborted 2025.04.01 17:45:29.378820 [ 40158 ] {} BaseDaemon: 2025.04.01 17:45:29.379100 [ 40158 ] {} BaseDaemon: Stack trace: 0x000055dfa744d11d 0x000055dfa784a422 0x000055df9eb397a6 0x000055df9eb39cfb 0x00007f1bbdcda520 0x00007f1bbdd2e9fd 0x00007f1bbdcda476 0x00007f1bbdcc07f3 0x000055df9eb38267 0x000055dfa7405a0a 0x000055dfa7406dff 0x000055df9ebbeade 0x000055df9ebce9ba 0x000055dfaf3e1a3c 0x000055dfaf3ded6e 0x000055dfaf3e0c0d 0x000055dfb3a56bfe 0x000055dfb3536f8b 0x000055dfb3a5684d 0x000055dfb355ba47 0x000055dfb354b54c 0x000055dfb354a8cc 0x000055dfb354a5aa 0x000055dfb354936a 0x000055dfa75187ee 0x000055dfa75209dc 0x000055df9eb31428 0x00007f1bbdd2cac3 0x00007f1bbddbe850 2025.04.01 17:45:29.379670 [ 40158 ] {} BaseDaemon: ######################################## 2025.04.01 17:45:29.380298 [ 40158 ] {} BaseDaemon: (version 24.12.2.20221.altinityantalya (altinity build), build id: 1D546DEA3F7DAB479A2081B78470431AC256AC96, git hash: 82252d159dc02cab0f366aaa5691adc1545dd11d) (from thread 2230) (query_id: 549c2d11-5172-44cf-9edd-d2571bb9f6a9) (query: INSERT INTO distributed_01099_b SELECT * FROM urlCluster('test_cluster_two_shards', 'http://localhost:8123/?query=select+{1,2,3}+format+TSV', 'TSV', 's String');) Received signal Aborted (6) 2025.04.01 17:45:29.381012 [ 40158 ] {} BaseDaemon: 2025.04.01 17:45:29.381547 [ 40158 ] {} BaseDaemon: Stack trace: 0x000055dfa744d11d 0x000055dfa784a422 0x000055df9eb397a6 0x000055df9eb39cfb 0x00007f1bbdcda520 0x00007f1bbdd2e9fd 0x00007f1bbdcda476 0x00007f1bbdcc07f3 0x000055df9eb38267 0x000055dfa7405a0a 0x000055dfa7406dff 0x000055df9ebbeade 0x000055df9ebce9ba 0x000055dfaf3e1a3c 0x000055dfaf3ded6e 0x000055dfaf3e0c0d 0x000055dfb3a56bfe 0x000055dfb3536f8b 0x000055dfb3a5684d 0x000055dfb355ba47 0x000055dfb354b54c 0x000055dfb354a8cc 0x000055dfb354a5aa 0x000055dfb354936a 0x000055dfa75187ee 0x000055dfa75209dc 0x000055df9eb31428 0x00007f1bbdd2cac3 0x00007f1bbddbe850 2025.04.01 17:45:29.582232 [ 40158 ] {} BaseDaemon: 0.0. inlined from ./build_docker/./src/Common/StackTrace.cpp:381: StackTrace::tryCapture() 2025.04.01 17:45:29.582934 [ 40158 ] {} BaseDaemon: 0. ./build_docker/./src/Common/StackTrace.cpp:350: StackTrace::StackTrace(ucontext_t const&) @ 0x000000001042411d 2025.04.01 17:45:29.745561 [ 40158 ] {} BaseDaemon: 1. ./build_docker/./src/Common/SignalHandlers.cpp:102: signalHandler(int, siginfo_t*, void*) @ 0x0000000010821422 2025.04.01 17:45:32.123765 [ 40158 ] {} BaseDaemon: 2. __tsan::CallUserSignalHandler(__tsan::ThreadState*, bool, bool, int, __sanitizer::__sanitizer_siginfo*, void*) @ 0x0000000007b107a6 2025.04.01 17:45:32.261573 [ 40158 ] {} BaseDaemon: 3. sighandler(int, __sanitizer::__sanitizer_siginfo*, void*) @ 0x0000000007b10cfb 2025.04.01 17:45:32.261814 [ 40158 ] {} BaseDaemon: 4. ? @ 0x00007f1bbdcda520 2025.04.01 17:45:32.262082 [ 40158 ] {} BaseDaemon: 5. ? @ 0x00007f1bbdd2e9fd 2025.04.01 17:45:32.262239 [ 40158 ] {} BaseDaemon: 6. ? @ 0x00007f1bbdcda476 2025.04.01 17:45:32.262493 [ 40158 ] {} BaseDaemon: 7. ? @ 0x00007f1bbdcc07f3 2025.04.01 17:45:32.409562 [ 40158 ] {} BaseDaemon: 8. ___interceptor_abort @ 0x0000000007b0f267 2025.04.01 17:45:32.688103 [ 40158 ] {} BaseDaemon: 9. ./build_docker/./src/Common/Exception.cpp:48: DB::abortOnFailedAssertion(String const&, void* const*, unsigned long, unsigned long) @ 0x00000000103dca0a 2025.04.01 17:45:32.974951 [ 40158 ] {} BaseDaemon: 10.0. inlined from ./build_docker/./src/Common/Exception.cpp:70: DB::handle_error_code(String const&, int, bool, std::vector> const&) 2025.04.01 17:45:32.975240 [ 40158 ] {} BaseDaemon: 10. ./build_docker/./src/Common/Exception.cpp:111: DB::Exception::Exception(DB::Exception::MessageMasked&&, int, bool) @ 0x00000000103dddff 2025.04.01 17:45:34.785308 [ 40158 ] {} BaseDaemon: 11. DB::Exception::Exception(PreformattedMessage&&, int) @ 0x0000000007b95ade 2025.04.01 17:45:34.920448 [ 40158 ] {} BaseDaemon: 12. DB::Exception::Exception<>(int, FormatStringHelperImpl<>) @ 0x0000000007ba59ba 2025.04.01 17:45:35.364440 [ 40158 ] {} BaseDaemon: 13. ./build_docker/./src/QueryPipeline/RemoteQueryExecutor.cpp:727: DB::RemoteQueryExecutor::processReadTaskRequest() @ 0x00000000183b8a3c 2025.04.01 17:45:35.710991 [ 40158 ] {} BaseDaemon: 14. ./build_docker/./src/QueryPipeline/RemoteQueryExecutor.cpp:623: DB::RemoteQueryExecutor::processPacket(DB::Packet) @ 0x00000000183b5d6e 2025.04.01 17:45:36.167573 [ 40158 ] {} BaseDaemon: 15. ./build_docker/./src/QueryPipeline/RemoteQueryExecutor.cpp:562: DB::RemoteQueryExecutor::readAsync() @ 0x00000000183b7c0d 2025.04.01 17:45:36.331098 [ 40158 ] {} BaseDaemon: 16. ./build_docker/./src/Processors/Sources/RemoteSource.cpp:182: DB::RemoteSource::tryGenerate() @ 0x000000001ca2dbfe 2025.04.01 17:45:36.417293 [ 40158 ] {} BaseDaemon: 17. ./build_docker/./src/Processors/ISource.cpp:108: DB::ISource::work() @ 0x000000001c50df8b 2025.04.01 17:45:36.515182 [ 40158 ] {} BaseDaemon: 18. ./build_docker/./src/Processors/Sources/RemoteSource.cpp:134: DB::RemoteSource::work() @ 0x000000001ca2d84d 2025.04.01 17:45:36.572278 [ 40158 ] {} BaseDaemon: 19.0. inlined from ./build_docker/./src/Processors/Executors/ExecutionThreadContext.cpp:49: DB::executeJob(DB::ExecutingGraph::Node*, DB::ReadProgressCallback*) 2025.04.01 17:45:36.572595 [ 40158 ] {} BaseDaemon: 19. ./build_docker/./src/Processors/Executors/ExecutionThreadContext.cpp:98: DB::ExecutionThreadContext::executeTask() @ 0x000000001c532a47 2025.04.01 17:45:36.688121 [ 40158 ] {} BaseDaemon: 20. ./build_docker/./src/Processors/Executors/PipelineExecutor.cpp:290: DB::PipelineExecutor::executeStepImpl(unsigned long, std::atomic*) @ 0x000000001c52254c 2025.04.01 17:45:36.790313 [ 40158 ] {} BaseDaemon: 21.0. inlined from ./build_docker/./src/Processors/Executors/PipelineExecutor.cpp:256: DB::PipelineExecutor::executeSingleThread(unsigned long) 2025.04.01 17:45:36.790775 [ 40158 ] {} BaseDaemon: 21. ./build_docker/./src/Processors/Executors/PipelineExecutor.cpp:442: DB::PipelineExecutor::executeImpl(unsigned long, bool) @ 0x000000001c5218cc 2025.04.01 17:45:36.901081 [ 40158 ] {} BaseDaemon: 22. ./build_docker/./src/Processors/Executors/PipelineExecutor.cpp:127: DB::PipelineExecutor::execute(unsigned long, bool) @ 0x000000001c5215aa 2025.04.01 17:45:36.972964 [ 40158 ] {} BaseDaemon: 23.0. inlined from ./build_docker/./src/Processors/Executors/CompletedPipelineExecutor.cpp:49: DB::threadFunction(DB::CompletedPipelineExecutor::Data&, std::shared_ptr, unsigned long, bool) 2025.04.01 17:45:36.973222 [ 40158 ] {} BaseDaemon: 23.1. inlined from ./build_docker/./src/Processors/Executors/CompletedPipelineExecutor.cpp:89: operator() 2025.04.01 17:45:36.973358 [ 40158 ] {} BaseDaemon: 23.2. inlined from ./contrib/llvm-project/libcxx/include/__functional/invoke.h:394: ? 2025.04.01 17:45:36.973992 [ 40158 ] {} BaseDaemon: 23.3. inlined from ./contrib/llvm-project/libcxx/include/tuple:1789: _ZNSt3__118__apply_tuple_implB6v15007IRZN2DB25CompletedPipelineExecutor7executeEvE3$_0RNS_5tupleIJEEETpTnmJEEEDcOT_OT0_NS_15__tuple_indicesIJXspT1_EEEE 2025.04.01 17:45:36.974829 [ 40158 ] {} BaseDaemon: 23.4. inlined from ./contrib/llvm-project/libcxx/include/tuple:1798: decltype(auto) std::apply[abi:v15007]&>(DB::CompletedPipelineExecutor::execute()::$_0&, std::tuple<>&) 2025.04.01 17:45:36.975028 [ 40158 ] {} BaseDaemon: 23.5. inlined from ./src/Common/ThreadPool.h:311: operator() 2025.04.01 17:45:36.975334 [ 40158 ] {} BaseDaemon: 23.6. inlined from ./contrib/llvm-project/libcxx/include/__functional/invoke.h:394: ? 2025.04.01 17:45:36.975646 [ 40158 ] {} BaseDaemon: 23.7. inlined from ./contrib/llvm-project/libcxx/include/__functional/invoke.h:479: ? 2025.04.01 17:45:36.975876 [ 40158 ] {} BaseDaemon: 23.8. inlined from ./contrib/llvm-project/libcxx/include/__functional/function.h:235: ? 2025.04.01 17:45:36.976343 [ 40158 ] {} BaseDaemon: 23. ./contrib/llvm-project/libcxx/include/__functional/function.h:716: ? @ 0x000000001c52036a 2025.04.01 17:45:37.070390 [ 40158 ] {} BaseDaemon: 24.0. inlined from ./contrib/llvm-project/libcxx/include/__functional/function.h:848: ? 2025.04.01 17:45:37.070675 [ 40158 ] {} BaseDaemon: 24.1. inlined from ./contrib/llvm-project/libcxx/include/__functional/function.h:1197: ? 2025.04.01 17:45:37.070894 [ 40158 ] {} BaseDaemon: 24. ./build_docker/./src/Common/ThreadPool.cpp:785: ThreadPoolImpl::ThreadFromThreadPool::worker() @ 0x00000000104ef7ee 2025.04.01 17:45:37.228767 [ 40158 ] {} BaseDaemon: 25.0. inlined from ./contrib/llvm-project/libcxx/include/__functional/invoke.h:359: ? 2025.04.01 17:45:37.229273 [ 40158 ] {} BaseDaemon: 25.1. inlined from ./contrib/llvm-project/libcxx/include/thread:284: void std::__thread_execute[abi:v15007]>, void (ThreadPoolImpl::ThreadFromThreadPool::*)(), ThreadPoolImpl::ThreadFromThreadPool*, 2ul>(std::tuple>, void (ThreadPoolImpl::ThreadFromThreadPool::*)(), ThreadPoolImpl::ThreadFromThreadPool*>&, std::__tuple_indices<2ul>) 2025.04.01 17:45:37.229646 [ 40158 ] {} BaseDaemon: 25. ./contrib/llvm-project/libcxx/include/thread:295: void* std::__thread_proxy[abi:v15007]>, void (ThreadPoolImpl::ThreadFromThreadPool::*)(), ThreadPoolImpl::ThreadFromThreadPool*>>(void*) @ 0x00000000104f79dc 2025.04.01 17:45:38.524260 [ 40158 ] {} BaseDaemon: 26. __tsan_thread_start_func @ 0x0000000007b08428 2025.04.01 17:45:38.524385 [ 40158 ] {} BaseDaemon: 27. ? @ 0x00007f1bbdd2cac3 2025.04.01 17:45:38.524515 [ 40158 ] {} BaseDaemon: 28. ? @ 0x00007f1bbddbe850 2025.04.01 17:45:41.785737 [ 40158 ] {} BaseDaemon: Integrity check of the executable successfully passed (checksum: CAD5C347B3B532C05434FCAB6F84FB12) 2025.04.01 17:45:49.494525 [ 40158 ] {} BaseDaemon: This ClickHouse version is not official and should be upgraded to the official build. 2025.04.01 17:45:49.496511 [ 40158 ] {} BaseDaemon: Changed settings: min_compress_block_size = 838722, max_compress_block_size = 1814485, max_block_size = 43866, min_external_table_block_size_bytes = 0, max_joined_block_size_rows = 36602, max_insert_threads = 3, max_threads = 1, max_parsing_threads = 1, max_read_buffer_size = 772261, connect_timeout_with_failover_ms = 2000, connect_timeout_with_failover_secure_ms = 3000, idle_connection_timeout = 36000, s3_max_get_rps = 1000000, s3_max_get_burst = 2000000, s3_max_put_rps = 1000000, s3_max_put_burst = 2000000, s3_check_objects_after_upload = true, use_uncompressed_cache = true, max_remote_read_network_bandwidth = 1000000000000, max_remote_write_network_bandwidth = 1000000000000, max_local_read_bandwidth = 1000000000000, max_local_write_bandwidth = 1000000000000, stream_like_engine_allow_direct_select = true, enable_multiple_prewhere_read_steps = false, replication_wait_for_inactive_replica_timeout = 30, compile_sort_description = false, group_by_two_level_threshold = 756728, group_by_two_level_threshold_bytes = 14099342, allow_nonconst_timezone_arguments = true, parallel_distributed_insert_select = 1, input_format_parallel_parsing = false, min_chunk_bytes_for_parallel_parsing = 9298309, merge_tree_coarse_index_granularity = 16, min_bytes_to_use_direct_io = 10737418240, min_bytes_to_use_mmap_io = 10737418240, log_queries = true, insert_quorum_timeout = 60000, merge_tree_read_split_ranges_into_intersecting_and_non_intersecting_injection_probability = 0.4000000059604645, http_response_buffer_size = 9705764, fsync_metadata = true, query_plan_join_swap_table = true, http_send_timeout = 60., http_receive_timeout = 60., use_index_for_in_with_subqueries_max_values = 1000000000, opentelemetry_start_trace_probability = 0.10000000149011612, max_rows_to_read = 20000000, max_bytes_to_read = 1000000000000, max_bytes_to_read_leaf = 1000000000000, max_rows_to_group_by = 10000000000, max_bytes_before_external_group_by = 10737418240, max_rows_to_sort = 10000000000, max_bytes_to_sort = 10000000000, prefer_external_sort_block_bytes = 100000000, max_bytes_before_external_sort = 10737418240, max_bytes_before_remerge_sort = 936605911, max_result_rows = 1000000000, max_result_bytes = 1000000000, max_execution_time = 600., max_execution_time_leaf = 600., max_execution_speed = 100000000000, max_execution_speed_bytes = 10000000000000, timeout_before_checking_execution_speed = 300., max_estimated_execution_time = 600., max_columns_to_read = 20000, max_temporary_columns = 20000, max_temporary_non_const_columns = 20000, max_rows_in_set = 10000000000, max_bytes_in_set = 10000000000, max_rows_in_join = 10000000000, max_bytes_in_join = 10000000000, cross_join_min_rows_to_compress = 1, cross_join_min_bytes_to_compress = 1, max_rows_to_transfer = 1000000000, max_bytes_to_transfer = 1000000000, max_rows_in_distinct = 10000000000, max_bytes_in_distinct = 10000000000, max_memory_usage = 5000000000, max_memory_usage_for_user = 32000000000, max_untracked_memory = 1048576, memory_profiler_step = 1048576, max_network_bandwidth = 100000000000, max_network_bytes = 1000000000000, max_network_bandwidth_for_user = 100000000000, max_network_bandwidth_for_all_users = 100000000000, max_temporary_data_on_disk_size_for_user = 100000000000, max_temporary_data_on_disk_size_for_query = 100000000000, max_backup_bandwidth = 100000000000, log_comment = '02221_parallel_replicas_bug.sh', send_logs_level = 'error', prefer_localhost_replica = false, optimize_read_in_order = false, optimize_aggregation_in_order = true, aggregation_in_order_max_block_bytes = 15394136, read_in_order_two_level_merge_threshold = 79, max_hyperscan_regexp_length = 1000000, max_hyperscan_regexp_total_length = 10000000, allow_introspection_functions = true, database_atomic_wait_for_drop_and_detach_synchronously = true, optimize_append_index = true, query_cache_max_size_in_bytes = 10000000, query_cache_max_entries = 100000, distributed_ddl_entry_format_version = 6, external_storage_max_read_rows = 10000000000, external_storage_max_read_bytes = 10000000000, local_filesystem_read_method = 'pread', merge_tree_min_bytes_per_task_for_remote_reading = 8388608, merge_tree_compact_parts_min_granules_to_multibuffer_read = 123, async_insert_busy_timeout_max_ms = 5000, enable_filesystem_cache = true, enable_filesystem_cache_on_write_operations = true, filesystem_cache_segments_batch_size = 5, use_page_cache_for_disks_without_file_cache = true, page_cache_inject_eviction = true, load_marks_asynchronously = true, allow_prefetched_read_pool_for_remote_filesystem = false, allow_prefetched_read_pool_for_local_filesystem = false, filesystem_prefetch_max_memory_usage = 33554432, filesystem_prefetches_limit = 10, max_streams_for_merge_tree_reading = 1000, optimize_sorting_by_input_stream_properties = false, insert_keeper_max_retries = 100, insert_keeper_retry_initial_backoff_ms = 1, insert_keeper_retry_max_backoff_ms = 10, insert_keeper_fault_injection_probability = 0.009999999776482582, optimize_distinct_in_order = false, allow_experimental_parallel_reading_from_replicas = 0, parallel_replicas_for_non_replicated_merge_tree = true, parallel_replicas_local_plan = false, session_timezone = 'Mexico/BajaSur', output_format_native_write_json_as_string = true 2025.04.01 17:45:53.785273 [ 423 ] {} Application: Child process was terminated by signal 6. + rg -A50 -Fa ============ /var/log/clickhouse-server/stderr.log + : + data_path_config=--path=/var/lib/clickhouse/ + zstd --threads=0 + [[ -n '' ]] + '[' 1 -ne 0 ']' + for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log + clickhouse-local --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 --path=/var/lib/clickhouse/ --only-system-tables --stacktrace -q 'select * from system.query_log format TSVWithNamesAndTypes' + zstd --threads=0 + [[ 0 -eq 1 ]] + [[ 0 -eq 1 ]] + for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log + clickhouse-local --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 --path=/var/lib/clickhouse/ --only-system-tables --stacktrace -q 'select * from system.zookeeper_log format TSVWithNamesAndTypes' + zstd --threads=0 + [[ 0 -eq 1 ]] + [[ 0 -eq 1 ]] + for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log + zstd --threads=0 + clickhouse-local --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 --path=/var/lib/clickhouse/ --only-system-tables --stacktrace -q 'select * from system.trace_log format TSVWithNamesAndTypes' + [[ 0 -eq 1 ]] + [[ 0 -eq 1 ]] + for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log + clickhouse-local --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 --path=/var/lib/clickhouse/ --only-system-tables --stacktrace -q 'select * from system.transactions_info_log format TSVWithNamesAndTypes' + zstd --threads=0 + [[ 0 -eq 1 ]] + [[ 0 -eq 1 ]] + for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log + zstd --threads=0 + clickhouse-local --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 --path=/var/lib/clickhouse/ --only-system-tables --stacktrace -q 'select * from system.metric_log format TSVWithNamesAndTypes' + [[ 0 -eq 1 ]] + [[ 0 -eq 1 ]] + for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log + zstd --threads=0 + clickhouse-local --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 --path=/var/lib/clickhouse/ --only-system-tables --stacktrace -q 'select * from system.blob_storage_log format TSVWithNamesAndTypes' + [[ 0 -eq 1 ]] + [[ 0 -eq 1 ]] + for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log + clickhouse-local --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 --path=/var/lib/clickhouse/ --only-system-tables --stacktrace -q 'select * from system.error_log format TSVWithNamesAndTypes' + zstd --threads=0 + [[ 0 -eq 1 ]] + [[ 0 -eq 1 ]] + for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log + clickhouse-local --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 --path=/var/lib/clickhouse/ --only-system-tables --stacktrace -q 'select * from system.query_metric_log format TSVWithNamesAndTypes' + zstd --threads=0 + [[ 0 -eq 1 ]] + [[ 0 -eq 1 ]] + for trace_type in CPU Memory Real + clickhouse-local --path=/var/lib/clickhouse/ --only-system-tables -q ' select arrayStringConcat((arrayMap(x -> concat(splitByChar('\''/'\'', addressToLine(x))[-1], '\''#'\'', demangle(addressToSymbol(x)) ), trace)), '\'';'\'') AS stack, count(*) AS samples from system.trace_log where trace_type = '\''CPU'\'' group by trace order by samples desc settings allow_introspection_functions = 1 format TabSeparated' + zstd --threads=0 + for trace_type in CPU Memory Real + clickhouse-local --path=/var/lib/clickhouse/ --only-system-tables -q ' select arrayStringConcat((arrayMap(x -> concat(splitByChar('\''/'\'', addressToLine(x))[-1], '\''#'\'', demangle(addressToSymbol(x)) ), trace)), '\'';'\'') AS stack, count(*) AS samples from system.trace_log where trace_type = '\''Memory'\'' group by trace order by samples desc settings allow_introspection_functions = 1 format TabSeparated' + zstd --threads=0 + for trace_type in CPU Memory Real + clickhouse-local --path=/var/lib/clickhouse/ --only-system-tables -q ' select arrayStringConcat((arrayMap(x -> concat(splitByChar('\''/'\'', addressToLine(x))[-1], '\''#'\'', demangle(addressToSymbol(x)) ), trace)), '\'';'\'') AS stack, count(*) AS samples from system.trace_log where trace_type = '\''Real'\'' group by trace order by samples desc settings allow_introspection_functions = 1 format TabSeparated' + zstd --threads=0 + check_logs_for_critical_errors + sed -n '/WARNING:.*anitizer/,/^$/p' /var/log/clickhouse-server/stderr.log + rg -Fav -e 'ASan doesn'\''t fully support makecontext/swapcontext functions' -e DB::Exception /test_output/tmp + echo -e 'No sanitizer asserts\tOK\t\N\t' + rm -f /test_output/tmp + rg -Fa ' Application: Child process was terminated by signal 9' /var/log/clickhouse-server/clickhouse-server.err.log /var/log/clickhouse-server/clickhouse-server.log + echo -e 'No OOM messages in clickhouse-server.log\tOK\t\N\t' + rg -Fa 'Code: 49. DB::Exception: ' /var/log/clickhouse-server/clickhouse-server.err.log /var/log/clickhouse-server/clickhouse-server.log + echo -e 'No logical errors\tOK\t\N\t' + '[' -s /test_output/logical_errors.txt ']' + rm /test_output/logical_errors.txt + rg --text 'Code: 499.*The specified key does not exist' /var/log/clickhouse-server/clickhouse-server.err.log /var/log/clickhouse-server/clickhouse-server.log + grep -v -e a.myext -e DistributedCacheTCPHandler -e ReadBufferFromDistributedCache -e ReadBufferFromS3 -e ReadBufferFromAzureBlobStorage -e AsynchronousBoundedReadBuffer -e 'caller id: None:DistribCache' + echo -e 'No lost s3 keys\tOK\t\N\t' + rg -Fa 'it is lost forever' /var/log/clickhouse-server/clickhouse-server.err.log /var/log/clickhouse-server/clickhouse-server.log + grep SharedMergeTreePartCheckThread + echo -e 'No SharedMergeTree lost forever in clickhouse-server.log\tOK\t\N\t' + '[' -s /test_output/no_such_key_errors.txt ']' + rm /test_output/no_such_key_errors.txt + rg -Fa '#######################################' /var/log/clickhouse-server/clickhouse-server.err.log /var/log/clickhouse-server/clickhouse-server.log + echo -e 'Killed by signal (in clickhouse-server.log)\tFAIL\t\N\t' + rg -Fa ' ' /var/log/clickhouse-server/clickhouse-server.err.log /var/log/clickhouse-server/clickhouse-server.log ++ trim_server_logs fatal_messages.txt ++ grep -Eo ' \[ [0-9]+ \] \{.*' ++ head -n 100 /test_output/fatal_messages.txt ++ escaped ++ clickhouse local -S 's String' --input-format=LineAsString -q 'select substr(s, 1, 300) from table format CustomSeparated settings format_custom_row_after_delimiter='\''\\\\n'\''' + echo -e 'Fatal message in clickhouse-server.log (see fatal_messages.txt)\tFAIL\t\N\t [ 2230 ] {549c2d11-5172-44cf-9edd-d2571bb9f6a9} : Logical error: \'\''Replica info is not initialized\'\''.\\n [ 2230 ] {549c2d11-5172-44cf-9edd-d2571bb9f6a9} : Stack trace (when copying this message, always include the lines below):\\n [ 40158 ] {} BaseDaemon: ########## Short fault info ############\\n [ 40158 ] {} BaseDaemon: (version 24.12.2.20221.altinityantalya (altinity build), build id: 1D546DEA3F7DAB479A2081B78470431AC256AC96, git hash: 82252d159dc02cab0f366aaa5691adc1545dd11d, architecture: x86_64) (from thread 2230) Received signal 6\\n [ 40158 ] {} BaseDaemon: Signal description: Aborted\\n [ 40158 ] {} BaseDaemon: \\n [ 40158 ] {} BaseDaemon: Stack trace: 0x000055dfa744d11d 0x000055dfa784a422 0x000055df9eb397a6 0x000055df9eb39cfb 0x00007f1bbdcda520 0x00007f1bbdd2e9fd 0x00007f1bbdcda476 0x00007f1bbdcc07f3 0x000055df9eb38267 0x000055dfa7405a0a 0x000055dfa7406dff 0x000055df9ebbeade 0x000055df9ebce9ba 0x0000\\n [ 40158 ] {} BaseDaemon: ########################################\\n [ 40158 ] {} BaseDaemon: (version 24.12.2.20221.altinityantalya (altinity build), build id: 1D546DEA3F7DAB479A2081B78470431AC256AC96, git hash: 82252d159dc02cab0f366aaa5691adc1545dd11d) (from thread 2230) (query_id: 549c2d11-5172-44cf-9edd-d2571bb9f6a9) (query: INSERT INTO distributed_01099\\n [ 40158 ] {} BaseDaemon: \\n [ 40158 ] {} BaseDaemon: Stack trace: 0x000055dfa744d11d 0x000055dfa784a422 0x000055df9eb397a6 0x000055df9eb39cfb 0x00007f1bbdcda520 0x00007f1bbdd2e9fd 0x00007f1bbdcda476 0x00007f1bbdcc07f3 0x000055df9eb38267 0x000055dfa7405a0a 0x000055dfa7406dff 0x000055df9ebbeade 0x000055df9ebce9ba 0x0000\\n [ 40158 ] {} BaseDaemon: 0.0. inlined from ./build_docker/./src/Common/StackTrace.cpp:381: StackTrace::tryCapture()\\n [ 40158 ] {} BaseDaemon: 0. ./build_docker/./src/Common/StackTrace.cpp:350: StackTrace::StackTrace(ucontext_t const&) @ 0x000000001042411d\\n [ 40158 ] {} BaseDaemon: 1. ./build_docker/./src/Common/SignalHandlers.cpp:102: signalHandler(int, siginfo_t*, void*) @ 0x0000000010821422\\n [ 40158 ] {} BaseDaemon: 2. __tsan::CallUserSignalHandler(__tsan::ThreadState*, bool, bool, int, __sanitizer::__sanitizer_siginfo*, void*) @ 0x0000000007b107a6\\n [ 40158 ] {} BaseDaemon: 3. sighandler(int, __sanitizer::__sanitizer_siginfo*, void*) @ 0x0000000007b10cfb\\n [ 40158 ] {} BaseDaemon: 4. ? @ 0x00007f1bbdcda520\\n [ 40158 ] {} BaseDaemon: 5. ? @ 0x00007f1bbdd2e9fd\\n [ 40158 ] {} BaseDaemon: 6. ? @ 0x00007f1bbdcda476\\n [ 40158 ] {} BaseDaemon: 7. ? @ 0x00007f1bbdcc07f3\\n [ 40158 ] {} BaseDaemon: 8. ___interceptor_abort @ 0x0000000007b0f267\\n [ 40158 ] {} BaseDaemon: 9. ./build_docker/./src/Common/Exception.cpp:48: DB::abortOnFailedAssertion(String const&, void* const*, unsigned long, unsigned long) @ 0x00000000103dca0a\\n [ 40158 ] {} BaseDaemon: 10.0. inlined from ./build_docker/./src/Common/Exception.cpp:70: DB::handle_error_code(String const&, int, bool, std::vector> const&)\\n [ 40158 ] {} BaseDaemon: 10. ./build_docker/./src/Common/Exception.cpp:111: DB::Exception::Exception(DB::Exception::MessageMasked&&, int, bool) @ 0x00000000103dddff\\n [ 40158 ] {} BaseDaemon: 11. DB::Exception::Exception(PreformattedMessage&&, int) @ 0x0000000007b95ade\\n [ 40158 ] {} BaseDaemon: 12. DB::Exception::Exception<>(int, FormatStringHelperImpl<>) @ 0x0000000007ba59ba\\n [ 40158 ] {} BaseDaemon: 13. ./build_docker/./src/QueryPipeline/RemoteQueryExecutor.cpp:727: DB::RemoteQueryExecutor::processReadTaskRequest() @ 0x00000000183b8a3c\\n [ 40158 ] {} BaseDaemon: 14. ./build_docker/./src/QueryPipeline/RemoteQueryExecutor.cpp:623: DB::RemoteQueryExecutor::processPacket(DB::Packet) @ 0x00000000183b5d6e\\n [ 40158 ] {} BaseDaemon: 15. ./build_docker/./src/QueryPipeline/RemoteQueryExecutor.cpp:562: DB::RemoteQueryExecutor::readAsync() @ 0x00000000183b7c0d\\n [ 40158 ] {} BaseDaemon: 16. ./build_docker/./src/Processors/Sources/RemoteSource.cpp:182: DB::RemoteSource::tryGenerate() @ 0x000000001ca2dbfe\\n [ 40158 ] {} BaseDaemon: 17. ./build_docker/./src/Processors/ISource.cpp:108: DB::ISource::work() @ 0x000000001c50df8b\\n [ 40158 ] {} BaseDaemon: 18. ./build_docker/./src/Processors/Sources/RemoteSource.cpp:134: DB::RemoteSource::work() @ 0x000000001ca2d84d\\n [ 40158 ] {} BaseDaemon: 19.0. inlined from ./build_docker/./src/Processors/Executors/ExecutionThreadContext.cpp:49: DB::executeJob(DB::ExecutingGraph::Node*, DB::ReadProgressCallback*)\\n [ 40158 ] {} BaseDaemon: 19. ./build_docker/./src/Processors/Executors/ExecutionThreadContext.cpp:98: DB::ExecutionThreadContext::executeTask() @ 0x000000001c532a47\\n [ 40158 ] {} BaseDaemon: 20. ./build_docker/./src/Processors/Executors/PipelineExecutor.cpp:290: DB::PipelineExecutor::executeStepImpl(unsigned long, std::atomic*) @ 0x000000001c52254c\\n [ 40158 ] {} BaseDaemon: 21.0. inlined from ./build_docker/./src/Processors/Executors/PipelineExecutor.cpp:256: DB::PipelineExecutor::executeSingleThread(unsigned long)\\n [ 40158 ] {} BaseDaemon: 21. ./build_docker/./src/Processors/Executors/PipelineExecutor.cpp:442: DB::PipelineExecutor::executeImpl(unsigned long, bool) @ 0x000000001c5218cc\\n [ 40158 ] {} BaseDaemon: 22. ./build_docker/./src/Processors/Executors/PipelineExecutor.cpp:127: DB::PipelineExecutor::execute(unsigned long, bool) @ 0x000000001c5215aa\\n [ 40158 ] {} BaseDaemon: 23.0. inlined from ./build_docker/./src/Processors/Executors/CompletedPipelineExecutor.cpp:49: DB::threadFunction(DB::CompletedPipelineExecutor::Data&, std::shared_ptr, unsigned long, bool)\\n [ 40158 ] {} BaseDaemon: 23.1. inlined from ./build_docker/./src/Processors/Executors/CompletedPipelineExecutor.cpp:89: operator()\\n [ 40158 ] {} BaseDaemon: 23.2. inlined from ./contrib/llvm-project/libcxx/include/__functional/invoke.h:394: ?\\n [ 40158 ] {} BaseDaemon: 23.3. inlined from ./contrib/llvm-project/libcxx/include/tuple:1789: _ZNSt3__118__apply_tuple_implB6v15007IRZN2DB25CompletedPipelineExecutor7executeEvE3$_0RNS_5tupleIJEEETpTnmJEEEDcOT_OT0_NS_15__tuple_indicesIJXspT1_EEEE\\n [ 40158 ] {} BaseDaemon: 23.4. inlined from ./contrib/llvm-project/libcxx/include/tuple:1798: decltype(auto) std::apply[abi:v15007]&>(DB::CompletedPipelineExecutor::execute()::$_0&, std::tuple<>&)\\n [ 40158 ] {} BaseDaemon: 23.5. inlined from ./src/Common/ThreadPool.h:311: operator()\\n [ 40158 ] {} BaseDaemon: 23.6. inlined from ./contrib/llvm-project/libcxx/include/__functional/invoke.h:394: ?\\n [ 40158 ] {} BaseDaemon: 23.7. inlined from ./contrib/llvm-project/libcxx/include/__functional/invoke.h:479: ?\\n [ 40158 ] {} BaseDaemon: 23.8. inlined from ./contrib/llvm-project/libcxx/include/__functional/function.h:235: ?\\n [ 40158 ] {} BaseDaemon: 23. ./contrib/llvm-project/libcxx/include/__functional/function.h:716: ? @ 0x000000001c52036a\\n [ 40158 ] {} BaseDaemon: 24.0. inlined from ./contrib/llvm-project/libcxx/include/__functional/function.h:848: ?\\n [ 40158 ] {} BaseDaemon: 24.1. inlined from ./contrib/llvm-project/libcxx/include/__functional/function.h:1197: ?\\n [ 40158 ] {} BaseDaemon: 24. ./build_docker/./src/Common/ThreadPool.cpp:785: ThreadPoolImpl::ThreadFromThreadPool::worker() @ 0x00000000104ef7ee\\n [ 40158 ] {} BaseDaemon: 25.0. inlined from ./contrib/llvm-project/libcxx/include/__functional/invoke.h:359: ?\\n [ 40158 ] {} BaseDaemon: 25.1. inlined from ./contrib/llvm-project/libcxx/include/thread:284: void std::__thread_execute[abi:v15007]>, void (ThreadPoolImpl::ThreadFromThreadPool::*)(), ThreadPoolImp\\n [ 40158 ] {} BaseDaemon: 25. ./contrib/llvm-project/libcxx/include/thread:295: void* std::__thread_proxy[abi:v15007]>, void (ThreadPoolImpl::ThreadFromThreadPool::*)(), ThreadPoolImpl BaseDaemon: 26. __tsan_thread_start_func @ 0x0000000007b08428\\n [ 40158 ] {} BaseDaemon: 27. ? @ 0x00007f1bbdd2cac3\\n [ 40158 ] {} BaseDaemon: 28. ? @ 0x00007f1bbddbe850\\n [ 40158 ] {} BaseDaemon: Integrity check of the executable successfully passed (checksum: CAD5C347B3B532C05434FCAB6F84FB12)\\n [ 40158 ] {} BaseDaemon: This ClickHouse version is not official and should be upgraded to the official build.\\n [ 40158 ] {} BaseDaemon: Changed settings: min_compress_block_size = 838722, max_compress_block_size = 1814485, max_block_size = 43866, min_external_table_block_size_bytes = 0, max_joined_block_size_rows = 36602, max_insert_threads = 3, max_threads = 1, max_parsing_threads = 1, max_read_buf\\n [ 423 ] {} Application: Child process was terminated by signal 6.\\n [ 2230 ] {549c2d11-5172-44cf-9edd-d2571bb9f6a9} : Logical error: \'\''Replica info is not initialized\'\''.\\n [ 2230 ] {549c2d11-5172-44cf-9edd-d2571bb9f6a9} : Stack trace (when copying this message, always include the lines below):\\n [ 40158 ] {} BaseDaemon: ########## Short fault info ############\\n [ 40158 ] {} BaseDaemon: (version 24.12.2.20221.altinityantalya (altinity build), build id: 1D546DEA3F7DAB479A2081B78470431AC256AC96, git hash: 82252d159dc02cab0f366aaa5691adc1545dd11d, architecture: x86_64) (from thread 2230) Received signal 6\\n [ 40158 ] {} BaseDaemon: Signal description: Aborted\\n [ 40158 ] {} BaseDaemon: \\n [ 40158 ] {} BaseDaemon: Stack trace: 0x000055dfa744d11d 0x000055dfa784a422 0x000055df9eb397a6 0x000055df9eb39cfb 0x00007f1bbdcda520 0x00007f1bbdd2e9fd 0x00007f1bbdcda476 0x00007f1bbdcc07f3 0x000055df9eb38267 0x000055dfa7405a0a 0x000055dfa7406dff 0x000055df9ebbeade 0x000055df9ebce9ba 0x0000\\n [ 40158 ] {} BaseDaemon: ########################################\\n [ 40158 ] {} BaseDaemon: (version 24.12.2.20221.altinityantalya (altinity build), build id: 1D546DEA3F7DAB479A2081B78470431AC256AC96, git hash: 82252d159dc02cab0f366aaa5691adc1545dd11d) (from thread 2230) (query_id: 549c2d11-5172-44cf-9edd-d2571bb9f6a9) (query: INSERT INTO distributed_01099\\n [ 40158 ] {} BaseDaemon: \\n [ 40158 ] {} BaseDaemon: Stack trace: 0x000055dfa744d11d 0x000055dfa784a422 0x000055df9eb397a6 0x000055df9eb39cfb 0x00007f1bbdcda520 0x00007f1bbdd2e9fd 0x00007f1bbdcda476 0x00007f1bbdcc07f3 0x000055df9eb38267 0x000055dfa7405a0a 0x000055dfa7406dff 0x000055df9ebbeade 0x000055df9ebce9ba 0x0000\\n [ 40158 ] {} BaseDaemon: 0.0. inlined from ./build_docker/./src/Common/StackTrace.cpp:381: StackTrace::tryCapture()\\n [ 40158 ] {} BaseDaemon: 0. ./build_docker/./src/Common/StackTrace.cpp:350: StackTrace::StackTrace(ucontext_t const&) @ 0x000000001042411d\\n [ 40158 ] {} BaseDaemon: 1. ./build_docker/./src/Common/SignalHandlers.cpp:102: signalHandler(int, siginfo_t*, void*) @ 0x0000000010821422\\n [ 40158 ] {} BaseDaemon: 2. __tsan::CallUserSignalHandler(__tsan::ThreadState*, bool, bool, int, __sanitizer::__sanitizer_siginfo*, void*) @ 0x0000000007b107a6\\n [ 40158 ] {} BaseDaemon: 3. sighandler(int, __sanitizer::__sanitizer_siginfo*, void*) @ 0x0000000007b10cfb\\n [ 40158 ] {} BaseDaemon: 4. ? @ 0x00007f1bbdcda520\\n [ 40158 ] {} BaseDaemon: 5. ? @ 0x00007f1bbdd2e9fd\\n [ 40158 ] {} BaseDaemon: 6. ? @ 0x00007f1bbdcda476\\n [ 40158 ] {} BaseDaemon: 7. ? @ 0x00007f1bbdcc07f3\\n [ 40158 ] {} BaseDaemon: 8. ___interceptor_abort @ 0x0000000007b0f267\\n [ 40158 ] {} BaseDaemon: 9. ./build_docker/./src/Common/Exception.cpp:48: DB::abortOnFailedAssertion(String const&, void* const*, unsigned long, unsigned long) @ 0x00000000103dca0a\\n [ 40158 ] {} BaseDaemon: 10.0. inlined from ./build_docker/./src/Common/Exception.cpp:70: DB::handle_error_code(String const&, int, bool, std::vector> const&)\\n [ 40158 ] {} BaseDaemon: 10. ./build_docker/./src/Common/Exception.cpp:111: DB::Exception::Exception(DB::Exception::MessageMasked&&, int, bool) @ 0x00000000103dddff\\n [ 40158 ] {} BaseDaemon: 11. DB::Exception::Exception(PreformattedMessage&&, int) @ 0x0000000007b95ade\\n [ 40158 ] {} BaseDaemon: 12. DB::Exception::Exception<>(int, FormatStringHelperImpl<>) @ 0x0000000007ba59ba\\n [ 40158 ] {} BaseDaemon: 13. ./build_docker/./src/QueryPipeline/RemoteQueryExecutor.cpp:727: DB::RemoteQueryExecutor::processReadTaskRequest() @ 0x00000000183b8a3c\\n [ 40158 ] {} BaseDaemon: 14. ./build_docker/./src/QueryPipeline/RemoteQueryExecutor.cpp:623: DB::RemoteQueryExecutor::processPacket(DB::Packet) @ 0x00000000183b5d6e\\n [ 40158 ] {} BaseDaemon: 15. ./build_docker/./src/QueryPipeline/RemoteQueryExecutor.cpp:562: DB::RemoteQueryExecutor::readAsync() @ 0x00000000183b7c0d\\n [ 40158 ] {} BaseDaemon: 16. ./build_docker/./src/Processors/Sources/RemoteSource.cpp:182: DB::RemoteSource::tryGenerate() @ 0x000000001ca2dbfe\\n [ 40158 ] {} BaseDaemon: 17. ./build_docker/./src/Processors/ISource.cpp:108: DB::ISource::work() @ 0x000000001c50df8b\\n [ 40158 ] {} BaseDaemon: 18. ./build_docker/./src/Processors/Sources/RemoteSource.cpp:134: DB::RemoteSource::work() @ 0x000000001ca2d84d\\n [ 40158 ] {} BaseDaemon: 19.0. inlined from ./build_docker/./src/Processors/Executors/ExecutionThreadContext.cpp:49: DB::executeJob(DB::ExecutingGraph::Node*, DB::ReadProgressCallback*)\\n [ 40158 ] {} BaseDaemon: 19. ./build_docker/./src/Processors/Executors/ExecutionThreadContext.cpp:98: DB::ExecutionThreadContext::executeTask() @ 0x000000001c532a47\\n [ 40158 ] {} BaseDaemon: 20. ./build_docker/./src/Processors/Executors/PipelineExecutor.cpp:290: DB::PipelineExecutor::executeStepImpl(unsigned long, std::atomic*) @ 0x000000001c52254c\\n [ 40158 ] {} BaseDaemon: 21.0. inlined from ./build_docker/./src/Processors/Executors/PipelineExecutor.cpp:256: DB::PipelineExecutor::executeSingleThread(unsigned long)\\n [ 40158 ] {} BaseDaemon: 21. ./build_docker/./src/Processors/Executors/PipelineExecutor.cpp:442: DB::PipelineExecutor::executeImpl(unsigned long, bool) @ 0x000000001c5218cc\\n [ 40158 ] {} BaseDaemon: 22. ./build_docker/./src/Processors/Executors/PipelineExecutor.cpp:127: DB::PipelineExecutor::execute(unsigned long, bool) @ 0x000000001c5215aa\\n [ 40158 ] {} BaseDaemon: 23.0. inlined from ./build_docker/./src/Processors/Executors/CompletedPipelineExecutor.cpp:49: DB::threadFunction(DB::CompletedPipelineExecutor::Data&, std::shared_ptr, unsigned long, bool)\\n' + '[' -s /test_output/fatal_messages.txt ']' + rg -v 'rg -Faz ' + rg -Faz '########################################' /test_output/blob_storage_log.tsv.zst /test_output/clickhouse-server.log.zst /test_output/error_log.tsv.zst /test_output/fatal_messages.txt /test_output/gdb.log /test_output/metric_log.tsv.zst /test_output/minio.log /test_output/query_log.tsv.zst /test_output/query_metric_log.tsv.zst /test_output/run.log /test_output/test_results.tsv /test_output/test_result.txt /test_output/trace-log-CPU-flamegraph.tsv.zst /test_output/trace-log-Memory-flamegraph.tsv.zst /test_output/trace-log-Real-flamegraph.tsv.zst /test_output/trace_log.tsv.zst /test_output/transactions_info_log.tsv.zst /test_output/zookeeper_log.tsv.zst + echo -e 'Killed by signal (output files)\tFAIL\t\N\t' + rg -Fa ' received signal ' /test_output/gdb.log + dmesg -T + grep -q -F -e 'Out of memory: Killed process' -e 'oom_reaper: reaped process' -e oom-kill:constraint=CONSTRAINT_NONE /test_output/dmesg.log + echo -e 'No OOM in dmesg\tOK\t\N\t' + /repo/tests/docker_scripts/process_functional_tests_result.py 2025-04-01 17:59:55,658 File /repo/tests/broken_tests.json with broken tests found 2025-04-01 17:59:55,658 Find files in result folder trace-log-Memory-flamegraph.tsv.zst,fatal_messages.txt,trace-log-CPU-flamegraph.tsv.zst,clickhouse-server.log.zst,error_log.tsv.zst,test_result.txt,zookeeper_log.tsv.zst,query_metric_log.tsv.zst,gdb.log,trace_log.tsv.zst,test_results.tsv,dmesg.log,run.log,minio.log,query_log.tsv.zst,trace-log-Real-flamegraph.tsv.zst,blob_storage_log.tsv.zst,metric_log.tsv.zst,transactions_info_log.tsv.zst 2025-04-01 17:59:55,662 Is flaky check: False 2025-04-01 17:59:55,662 Result parsed 2025-04-01 17:59:55,664 Result written File /repo/tests/analyzer_tech_debt.txt with broken tests found Broken tests in the list: 3 Broken tests in the list: 3 + rm /var/log/clickhouse-server/clickhouse-server.log + mv /var/log/clickhouse-server/stderr.log /test_output/ + [[ -n '' ]] + tar -chf /test_output/coordination.tar /var/lib/clickhouse/coordination tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets + rm -rf /var/lib/clickhouse/data/system/asynchronous_insert_log/ /var/lib/clickhouse/data/system/asynchronous_metric_log/ /var/lib/clickhouse/data/system/backup_log/ /var/lib/clickhouse/data/system/blob_storage_log/ /var/lib/clickhouse/data/system/crash_log/ /var/lib/clickhouse/data/system/error_log/ /var/lib/clickhouse/data/system/filesystem_cache_log/ /var/lib/clickhouse/data/system/metric_log/ /var/lib/clickhouse/data/system/opentelemetry_span_log/ /var/lib/clickhouse/data/system/part_log/ /var/lib/clickhouse/data/system/processors_profile_log/ /var/lib/clickhouse/data/system/query_log/ /var/lib/clickhouse/data/system/query_metric_log/ /var/lib/clickhouse/data/system/query_thread_log/ /var/lib/clickhouse/data/system/query_views_log/ /var/lib/clickhouse/data/system/s3queue_log/ /var/lib/clickhouse/data/system/session_log/ /var/lib/clickhouse/data/system/text_log/ /var/lib/clickhouse/data/system/trace_log/ /var/lib/clickhouse/data/system/transactions_info_log/ /var/lib/clickhouse/data/system/zookeeper_log/ + tar -chf /test_output/store.tar /var/lib/clickhouse/store tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets + tar -chf /test_output/metadata.tar /var/lib/clickhouse/metadata/default.sql /var/lib/clickhouse/metadata/information_schema.sql /var/lib/clickhouse/metadata/INFORMATION_SCHEMA.sql /var/lib/clickhouse/metadata/system.sql /var/lib/clickhouse/metadata/test_89ntz7bn.sql /var/lib/clickhouse/metadata/test_axigm1xv.sql /var/lib/clickhouse/metadata/test_jypz7w9c.sql /var/lib/clickhouse/metadata/test_l0nrtmbt.sql /var/lib/clickhouse/metadata/test_ofjicghs.sql /var/lib/clickhouse/metadata/test_ofr5rq7t.sql /var/lib/clickhouse/metadata/test_rnvzzqsk.sql /var/lib/clickhouse/metadata/test_sczjo540.sql /var/lib/clickhouse/metadata/test.sql /var/lib/clickhouse/metadata/test_umfejymz.sql tar: Removing leading `/' from member names tar: Removing leading `/' from hard link targets + [[ 0 -eq 1 ]] + [[ 0 -eq 1 ]] + collect_core_dumps + find . -type f -maxdepth 1 -name 'core.*' + read -r core